Skip to main content
Log in

Effective use of analysts’ effort in automated tracing

  • Original Article
  • Published:
Requirements Engineering Aims and scope Submit manuscript

Abstract

Because of the large amount of effort it takes to manually trace requirements, automated traceability methods for mapping textual software engineering artifacts to each other and generating candidate links have received increased attention over the past 15 years. Automatically generated links, however, are viewed as candidates until human analysts confirm/reject them for the final requirements traceability matrix. Studies have shown that analysts are a fallible, but necessary, participant in the tracing process. There are two key measures guiding analyst work on the evaluation of candidate links: accuracy of analyst decision and efficiency of their work. Intuitively, it is expected that the more effort the analyst spends on candidate link validation, the more accurate the final traceability matrix is likely to be, although the exact nature of this relationship may be difficult to gauge outright. To assist analysts in making the best use of their time when reviewing candidate links, prior work simulated four possible behaviors and showed that more structured approaches save the analysts’ time/effort required to achieve certain levels of accuracy. However, these behavioral simulations are complex to run and their results difficult to interpret and use in practice. In this paper, we present a mathematical model for evaluating analyst effort during requirements tracing tasks. We apply this model to a simulation study of 12 candidate link validation approaches. The simulation study is conducted on a number of different datasets. In each study, we assume perfect analyst behavior (i.e., analyst always being correct when making a determination about a link). Under this assumption, we evaluate the estimated effort for the analyst and plot it against the accuracy of the recovered traceability matrix. The effort estimation model is guided by a parameter specifying the relationship between the time it takes an analyst to evaluate a presented link and the time it takes an analyst to discover a link not presented to her. We construct a series of effort estimations based on different values of the model parameter. We found that the analysts’ approach to candidate link validation—essentially the order in which the analyst examines presented candidate links—does impact the effort. We also found that the lowest ratio of the cost of finding a correct link from scratch over the cost of recognizing a correct link yields the lowest effort for all datasets, but that the lowest effort does not always yield the highest quality matrix. We finally observed that effort varies by dataset. We conclude that the link evaluation approach we call “Top 1 Not Yet Examined Feedback Pruning” was the overall winner in terms of effort and highest quality and, thus, should be followed by human analysts if possible.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13
Fig. 14
Fig. 15
Fig. 16
Fig. 17
Fig. 18
Fig. 19
Fig. 20
Fig. 21
Fig. 22

Similar content being viewed by others

Notes

  1. Making the best use of analysts’ effort is difficult, which is the reason why “Cost-Effective” was named as the #2 Grand Challenge [4] facing requirement practitioners and researchers. The challenge reads “2. Cost-Effective—The return from using traceability is adequate in relation to the outlay of establishing it.”

  2. An automated method returns a link if it evaluates the similarity of the high-level and low-level elements forming the link above a certain, pre-defined in the method threshold value.

  3. These parameters were labeled α, β, and γ in prior work [5]. We rename them for this paper to avoid confusion with the Rochio feedback parameters.

  4. Note that “requirements tracing” is often the moniker even when requirements are not being traced.

  5. It should be noted that many of the 48 scenarios resulted in ties for minimum effort value; thus, there were far less than 48 scenarios that were “in the running” for best and worst.

References

  1. Gotel O, Finkelstein A (1997) Extended requirements traceability: results of an industrial case study. In: Proceedings of the 3rd IEEE international symposium on requirements engineering (RE’97). IEEE Computer Society, p 169

  2. Hayes JH, Dekhtyar A, Sundaram S (2006) Advancing candidate link generation for requirements tracing: the study of methods. IEEE Trans Softw Eng 32:4–19

    Article  Google Scholar 

  3. U.S. Department of Health and Human Services, Food and Drug Administration (2002) General principles of software validation; Final Guidance for Industry and FDA Staff

  4. Gotel O, Huang JC, Hayes JH, Zisman A, Egyed A, Grünbacher P, Dekhtyar A, Antoniol G, Maletic J, Mäder P (2012) Traceability fundamentals. In: Huang JC, Gotel O, Zisman A (eds) Software and systems traceability. Springer. ISBN: 1447122380

  5. Dekhtyar A, Hayes JH, Smith M (2011) Towards a model of analyst effort for traceability research: a position paper. In: Proceedings of traceability of emerging forms of software engineering (TEFSE)

  6. Hayes JH, Dekhtyar A, Osborne J (2003) Improving requirements tracing via information retrieval. In: International conference on requirements engineering, Monterey, California, pp 151–161

  7. Sundaram S, Hayes J, Dekhtyar A, Holbrook E (2010) Assessing traceability of software engineering artifacts. Requir Eng 15:313–335

    Article  Google Scholar 

  8. Axiom. http://www.iconcur-software.com/. Last accessed 29 June 2016

  9. Baeza-Yates R, Ribeiro-Neto B (1999) Modern information retrieval. ACM Press, Addison-Wesley

    Google Scholar 

  10. Hayes JH, Dekhtyar A, Sundaram S (2005) Improving after the fact tracing and mapping to support software quality predictions. IEEE Softw 22:30–37

    Article  Google Scholar 

  11. Antoniol G, Canfora G, Casazza G, Lucia AD, Merlo E (2002) Recovering traceability links between code and documentation. IEEE Trans Softw Eng 28:970–983

    Article  Google Scholar 

  12. Antoniol G, Caprile B, Potrich A, Tonella P (1999) Design-code traceability for object oriented systems. Ann Softw Eng 9:35–58

    Article  MATH  Google Scholar 

  13. Marcus A, Maletic J (2003) Recovering documentation-to-source code traceability links using latent semantic indexing. In: Proceedings of the twenty-fifth international conference on software engineering 2003, pp 125–135

  14. Hayes JH, Dekhtyar A (2005) Humans in the traceability loop: can’t live with ‘em, can’t live without ‘em. In: Proceedings of the 3rd international workshop on traceability in emerging forms of software engineering. ACM, New York, NY, USA, pp 20–23

  15. Cuddeback D, Dekhtyar A, Hayes J (2010) Automated requirements traceability: the study of human analysts. In: Requirements engineering, IEEE international conference on, vol 0, 2010, pp 231–240

  16. Dekhtyar A, Hayes JH, Larsen J (2007) Make the most of your time: how should the analyst work with automated traceability tools? In: Proceedings of the third international workshop on predictor models in software engineering. IEEE Computer Society, Washington, DC, USA

  17. Kong W-K, Hayes JH, Dekhtyar A, Dekhtyar O (2012) Process improvement for traceability: a study of human fallibility. In: Requirements engineering conference (RE), 2012 20th IEEE international, pp 31–40, 24–28 Sept. 2012

  18. Ali Nasir, Guéhéneuc Yann-Gaël, Antoniol Giuliano (2013) Trustrace: mining software repositories to improve the accuracy of requirement traceability links. Trans Softw Eng (TSE) 39(5):725–741

    Article  Google Scholar 

  19. “GanttProject Home”

  20. G. SBRS MODIS science data processing software requirements specification version 2

  21. Cleland-Huang J, Chang C, Sethi G, Javvaji K, Hu H, Xia J (2002) Automating speculative queries through event-based requirements traceability. In: Proceedings of the IEEE joint international requirements engineering conference (RE ‘02), 2002, pp 289–296

  22. Hayes JH, Dekhtyar A, Sundaram S, Howard S (2004) Helping analysts trace requirements: an objective look. In: International conference on requirements engineering (RE’2004)

  23. Hayes J, Dekhtyar A, Sundaram S, Holbrook E, Vadlamudi S, April A (2007) REquirements TRacing On target (RETRO): improving software maintenance through traceability recovery. Innov Syst Softw Eng 3:193–202

    Article  Google Scholar 

  24. Egyed A, Graf F, Grunbacher P (2010) Effort and quality of recovering requirements-to-code traces: two exploratory experiments. In: Requirements engineering, IEEE international conference on, vol 0, 2010, pp 221–230

  25. Zhang Z, Liu C, Zhang Y, Zhou T (2010) Solving the cold-start problem in recommender systems with social tags. CoRR, abs/1004.3732

  26. Lika B, Kolomvatsos K, Hadjiefthymiades S (2014) Facing the cold start problem in recommender systems. Expert Syst Appl 41(4, Part 2):2065–2073

    Article  Google Scholar 

  27. Bostandjiev S, O’Donovan J, Höllerer T (2012) TasteWeights: a visual interactive hybrid recommender system. In: Proceedings of the sixth ACM conference on recommender systems (RecSys ‘12). ACM, New York, NY, USA, pp 35–42

  28. Hariri N, Mobasher B Burke R (2014) Context adaptation in interactive recommender systems. In: Proceedings of the 8th ACM conference on recommender systems(RecSys ‘14). ACM, New York, NY, USA, pp 41–48

  29. CM1 DataSet, Metrics data program website, CM-1 project

Download references

Acknowledgements

We thank Matt Smith and Chelsea Hayes for their work on the spreadsheets and charts and tables in this paper. This work was partially sponsored by NASA under Grant NNX06AD02G. This work was funded in part by the National Science Foundation under NSF Grant CCF 0811140 and by a Lockheed Martin grant.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Jane Huffman Hayes.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Hayes, J.H., Dekhtyar, A., Larsen, J. et al. Effective use of analysts’ effort in automated tracing. Requirements Eng 23, 119–143 (2018). https://doi.org/10.1007/s00766-016-0260-8

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00766-016-0260-8

Keywords

Navigation