Sprint Management in Agile Approach: Progress and Velocity Evaluation Applying Machine Learning
"> Figure 1
<p>Burndown chart (in hours).</p> "> Figure 2
<p>Completed vs. not completed work (in hours).</p> "> Figure 3
<p>Access to the API method for obtaining the burndown chart.</p> "> Figure 4
<p>Access to the API Method for the completed vs. not completed UTs chart.</p> ">
Abstract
:1. Introduction
2. Related Works
3. Definition of the Problem
4. Development
- Training in the necessary knowledge (Kanban, Lean Development, Scrum, and Extreme Programming).
- Diagnosis of the area where the agile transformation will take place.
- Supporting tools for the approach. It includes proprietary tools: Agilev-Roadmap [20] for diagnosing and supporting the management of an agile roadmap and Worki, a tool for agile work management.
4.1. Sprint Evaluation in Worki
- Backlog Management and Product Structure: The project backlog management and the assessment of how the product is structured in the current Sprint.
- Project and Sprint Scope Management: An Analysis is conducted on how the project scope has been managed about the Sprint and whether the established objectives have been met.
- Time Invested, Estimates, and Re-estimations: Consideration is given to the time spent on the Sprint, initial estimates, and any re-estimations made during development.
- Progress, Work Pace, and Sprint Completion: Evaluating progress and work pace provides a clear view of Sprint’s status and helps identify potential issues. Completing Sprint’s work is a crucial objective.
- Preparation for the Next Sprint: Observation is made on how the next Sprint is being prepared and whether lessons learned are being applied.
- Product Owner Satisfaction: The satisfaction of the Product Owner is an essential indicator of Sprint’s quality and success.
- Burndown Chart: The goal is to reduce the remaining effort continuously.
- Completed and Not Completed Work Chart: This shows the percentage of work completed and not completed. This chart expects to show an increasing and continuous amount of completed work.
- Maximum Time in Activity Chart: Observes unattended work units in an activity for many days.
4.2. Research Questions
- How can ML techniques improve the accuracy and reliability of Sprint progress and work pace evaluations?
- What indicators, extracted from the Worki tool, most effectively contribute to predicting potential issues or delays in Sprints?
- Can ML models significantly improve detection Sprints that are falling behind compared to traditional evaluation methods?
- What are the best practices for integrating ML models into agile project management tools like Worki to enhance decision-making?
4.3. Dataset Design
4.4. Data Collection and Extraction Process
- Burndown chart data: The program retrieves data from the burndown chart via the Worki API, representing the remaining and recorded work overtime during the Sprint. The program uses these data to calculate Sprint’s progress and detect possible excessive reductions or lack of work pace.
- Percentage of completed work: The program calculates the percentage of completed work based on estimated completed and not completed work units. This provides a more rigorous measure of Sprint progress since each work unit is counted as either completed or not.
- Work reduction pace: The pace at which remaining work is being reduced in the Sprint is analyzed. The program identifies excessive reduction paces or the lack of significant reduction pace.
- Progressive increase in work: The program verifies a progressive increase in the remaining effort over time. If the remaining effort is not considerably decreasing, it is considered a sign of a lack of significant progress.
- Days of work stoppage: The number of days without significant progress in reducing the remaining effort is counted. This feature indicates the number of days the Sprint may have been stalled.
4.5. The Final Dataset
4.6. Data Analysis Methods
- Precision: This metric reflects the proportion of correct, optimistic predictions. In Sprint evaluation, high precision (as seen in the SVM, LR, and RF models) indicates that the model effectively identifies performance categories while avoiding false positives. For example, if a sprint is classified as Excellent, high precision suggests that the model accurately predicts this Sprint category.
- Recall: Recall measures the model’s ability to identify all positive instances correctly. High recall values, as exhibited by the SVM and LR models, are essential for identifying low-performing Sprints (Poor). This ensures that underperforming Sprints are captured accurately, aiding in better planning and optimizing future Sprints.
- F1 Score: The F1 score balances precision and recall. A high F1 score, seen in the SVM and LR models, suggests that these models can correctly identify performance categories while minimizing errors (false positives and negatives). This balance is vital for accurately assessing team performance in Sprints without over- or underestimating their effectiveness.
- Accuracy: Accuracy measures the proportion of correct predictions overall. In this case, the SVM, LR, and RF models, with accuracy close to or above 0.85, indicate a robust classification of Sprint performance categories. This implies that these models can correctly classify most Sprints into their respective categories, helping teams track progress more effectively.
- Correct predictions: The matrix diagonal represents the correct predictions, where instances are classified into the proper categories. The model classified 129 instances correctly in the Poor category, accurately identifying low-performance Sprints, which is crucial for Sprint evaluation. The SVM model correctly classified 62 instances for the Fair category, but some misclassifications need attention. The Good and Excellent categories saw high correct predictions, with 158 and 134 instances correctly classified, respectively.
- Misclassifications (off-diagonal elements): For Fair category, 16 instances labeled as Fair were incorrectly classified as Poor. This indicates the model struggles to distinguish some Fair Sprints from Poor ones, leading to false negatives. Twenty-one instances labeled as Fair were classified as Good, showing some overlap between the performance levels of these categories. For the Good category, five instances labeled as Fair were misclassified as Good, while four instances labeled Excellent were misclassified as Good. This could indicate challenges in recognizing nuances between high-performing Sprints. For the Excellent category: Only four instances of Excellent were incorrectly classified as Good, indicating strong model performance in identifying top-performing Sprints.
- The model performs well in correctly classifying Poor, Good, and Excellent Sprints, which is critical for identifying low- and high-performing sprints.
- The misclassification of some Fair instances, mainly being confused with Poor or Good, suggests that the model struggles slightly with mid-range performance levels, where the distinctions between categories are subtler.
5. Discussion and Analysis of Results
6. Conclusions
7. Future Work
- Improving the dataset: The idea of integrating additional sections of the Sprint evaluation from Worki is considered to strengthen the proposed dataset. This would involve incorporating other tracking charts available in the Worki tool.
- Model refinement: Conducting experiments with normalized numerical values, using feature weights and possibly including other supporting charts from Worki for tracking. This refinement aims to enhance the model’s performance and accuracy.
- Development of support tools: Creating tools and plugins that facilitate the implementation and use of the model in various platforms and development environments. For example, integrating the model with the Worki platform could provide detailed explanations of the evaluations made by the AI model.
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Sutherland, J. Manifesto for Agile Software Development. Available online: http://agilemanifesto.org/ (accessed on 8 January 2024).
- Digital.ai. 16th State of Agile Report: Agile Adoption Accelerates Across the Enterprise. Available online: https://stateofagile.com/ (accessed on 16 December 2023).
- Project Management Institute. Agile Practice Guide; Project Management Institute: Newton Square, PA, USA, 2017. [Google Scholar]
- Project Management Institute. A Guide to the Project Management Body of Knowledge (PMBOK® Guide), 6th ed.; Project Management Institute: Newtown Square, PA, USA, 2017. [Google Scholar]
- Schwaber, K.; Sutherland, J. La Guía Scrum. La Guía Definitiva de Scrum: Las Reglas del Juego. 2020. Available online: https://repositorio.uvm.edu.ve/server/api/core/bitstreams/5b4aef9b-52f7-49c5-8875-612b1b1dcbc0/content (accessed on 2 August 2022).
- Jadhav, A.; Kaur, M.; Akter, F. Evolution of Software Development Effort and Cost Estimation Techniques: Five Decades Study Using Automated Text Mining Approach. Math. Probl. Eng. 2022, 2022, 5782587. [Google Scholar] [CrossRef]
- Shamshurin, I.; Saltz, J.S. A predictive model to identify at-risk Kanban teams. Model Assist. Stat. Appl. 2019, 14, 321–335. [Google Scholar] [CrossRef]
- Sousa, A.; Faria, J.P.; Mendes-Moreira, J.; Gomes, D.; Henriques, P.C.; Graça, R. Applying Machine Learning to Risk Assessment in Software Projects. In Communications in Computer and Information Science (CCIS); Springer: Berlin/Heidelberg, Germany, 2021; Volume 1525, pp. 104–118. [Google Scholar] [CrossRef]
- Mashkoor, A.; Menzies, T.; Egyed, A.; Ramler, R. Artificial Intelligence and Software Engineering: Are We Ready? Computer 2022, 55, 24–28. [Google Scholar] [CrossRef]
- Abadeer, M.; Sabetzadeh, M. Machine Learning-based Estimation of Story Points in Agile Development: Industrial Experience and Lessons Learned. In Proceedings of the IEEE 29th International Requirements Engineering Conference Workshops (REW), Notre Dame, IN, USA, 20–24 September 2021; pp. 106–115. [Google Scholar] [CrossRef]
- Choetkiertikul, M.; Dam, H.K.; Tran, T.; Pham, T.; Ghose, A.; Menzies, T. A Deep Learning Model for Estimating Story Points. IEEE Trans. Softw. Eng. 2019, 45, 637–656. [Google Scholar] [CrossRef]
- Ramessur, M.A.; Nagowah, S.D. A predictive model to estimate effort in a sprint using machine learning techniques. Int. J. Inf. Technol. 2021, 13, 1101–1110. [Google Scholar] [CrossRef]
- Periyasamy, K.; Chianelli, J. A Project Tracking Tool for Scrum Projects with Machine Learning Support for Cost Estimation. EPIC Ser. Comput. 2021, 76, 86–94. [Google Scholar] [CrossRef]
- Malgonde, O.; Chari, K. An ensemble-based model for predicting agile software development effort. Empir. Softw. Eng. 2019, 24, 1017–1055. [Google Scholar] [CrossRef]
- Zia, Z.; Kamal, T.; Ziauddin, Z. An Effort Estimation Model for Agile Software Development. Adv. Comput. Sci. Its Appl. (ACSA) 2012, 2, 314–324. [Google Scholar]
- Batarseh, F.A.; Gonzalez, A.J. Predicting failures in agile software development through data analytics. Softw. Qual. J. 2018, 26, 49–66. [Google Scholar] [CrossRef]
- Dragicevic, S.; Celar, S.; Turic, M. Bayesian network model for task effort estimation in agile software development. J. Syst. Softw. 2017, 127, 109–119. [Google Scholar] [CrossRef]
- Srivastava, P.; Srivastava, N.; Agarwal, R.; Singh, P. Estimation in Agile Software Development Using Artificial Intelligence. In Lecture Notes in Networks and Systems; Springer: Berlin/Heidelberg, Germany, 2022; Volume 376, pp. 83–93. [Google Scholar] [CrossRef]
- Atlassian. Jira. Available online: https://www.atlassian.com/es/software/jira (accessed on 8 January 2024).
- TUNE-UP Process. Available online: http://www.tuneupprocess.com/ (accessed on 8 January 2024).
Key | Name | Description |
---|---|---|
R1 | Percentage of Work Completed | Percentage of work completed by the end of the Sprint. |
R2 | No Significant Progress? | Are there periods when there are no new completed work? |
R3 | Times Without Significant Progress | Number of periods without new completed work. |
R4 | Percentage of Work Stalled | % of days with unfinished work relative to the total days of the Sprint. |
R5 | Total Days Without Progress | Total days with unfinished work. |
R6 | No Significant Reduction Pace? | There was no significant reduction in the remaining programming effort between two consecutive days. |
R7 | Count of Days Without Significant Reduction | Number of days without significant reduction in the remaining programming effort. |
R8 | Excessive Reduction? | Excessive reduction in the remaining programming effort between two days. |
R9 | Count of Excessive Reductions | Several excessive reductions in the remaining programming effort between two days. Causes: overestimating task units (fewer programming hours needed) or scope reduction during the Sprint. |
R10 | Significant Increase? | Increase the remaining programming effort between two days. |
R11 | Count of Significant Increases | Several significant increases were in the remaining programming effort between the two days. Causes: increase in estimation or introduction of new task units during the Sprint. |
R1 | R2 | R3 | R4 | R5 | R6 | R7 | R8 | R9 | R10 | R11 | TAG |
---|---|---|---|---|---|---|---|---|---|---|---|
100 | 1 | 1 | 33.33 | 7 | 1 | 1 | 0 | 0 | 0 | 0 | 2 |
100 | 1 | 1 | 47.62 | 10 | 0 | 0 | 0 | 0 | 0 | 0 | 2 |
65.84 | 1 | 1 | 66.67 | 14 | 1 | 2 | 1 | 1 | 0 | 0 | 1 |
Recall | F1 | Accuracy | Precision | |
---|---|---|---|---|
KNN | 0.86 | 0.86 | 0.86 | 0.87 |
SVM | 0.88 | 0.87 | 0.88 | 0.88 |
MLP | 0.86 | 0.86 | 0.86 | 0.87 |
RF | 0.87 | 0.86 | 0.87 | 0.87 |
NB | 0.69 | 0.66 | 0.69 | 0.78 |
LR | 0.87 | 0.87 | 0.87 | 0.88 |
AB | 0.77 | 0.75 | 0.77 | 0.75 |
Poor | Fair | Good | Excellent | |
---|---|---|---|---|
Poor | 129 | 15 | 2 | 0 |
Fair | 16 | 62 | 21 | 1 |
Good | 0 | 5 | 158 | 4 |
Excellent | 0 | 0 | 4 | 134 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Pérez Castillo, Y.J.; Orantes Jiménez, S.D.; Letelier Torres, P.O. Sprint Management in Agile Approach: Progress and Velocity Evaluation Applying Machine Learning. Information 2024, 15, 726. https://doi.org/10.3390/info15110726
Pérez Castillo YJ, Orantes Jiménez SD, Letelier Torres PO. Sprint Management in Agile Approach: Progress and Velocity Evaluation Applying Machine Learning. Information. 2024; 15(11):726. https://doi.org/10.3390/info15110726
Chicago/Turabian StylePérez Castillo, Yadira Jazmín, Sandra Dinora Orantes Jiménez, and Patricio Orlando Letelier Torres. 2024. "Sprint Management in Agile Approach: Progress and Velocity Evaluation Applying Machine Learning" Information 15, no. 11: 726. https://doi.org/10.3390/info15110726