Abstract
Thales Alenia Space has recently widely deployed user centered design process in their software product conception. This standard breaks today technology and data centric approach by integrating end-users all along the iterative design stages: context and usage understanding, end-user need specification, quick mockups and end-user validation. This paper is a return on experience. It describes and investigates a dirty UCD methodology relevance based on prototyping and user testing only (skipping user research first activities). This process is made to fit project which needs front-end requirements at day one.
You have full access to this open access chapter, Download conference paper PDF
Similar content being viewed by others
Keywords
1 Introduction
User centered design is a well-known software conception standard in ergonomics and human factors fields (ISO 9241-210) [1]. To guarantee system’s utility and usability, it implies four stages where end-user is integrated within many ergonomics methods such as focus group [2], participatory design [3], interviews, job task-model based analysis [4], persona specification [5] and so on. In some constrained industrial context, deploying every practices cannot be possible. Thales Alenia Space believes that every process can be tailored and scaled to fit every project. For our new satellite command center (SCC) procedure executor named PRISM, we try a dirty UCD approach based on incremental mockups and user testing only. As user requirements were taken from our previous product version, this process enables earlier front-end development than classical ergonomics methodology.
This testimonial explains first how we integrated UCD activities within agile development process. Then we briefly present our study case PRISM and its incremental mockups. Next, we inspect our process relevance from usability metrics and user feedbacks. Does this approach lead to solution convergence? Does it bring high quality interaction and high user satisfaction? Does it allow continuous interaction improvement? From these problematic, the paper concludes on a preliminary analysis of this dirty UCD approach (ongoing work). Please note this article doesn’t list any SCC user needs. It only aims to analyze the process relevance from ergonomics point of view.
2 Process Overview
Industry often works with subcontractors while developing products. When interaction designers are not involved in project planning phase, it may results to a development-first approach without much consideration for ergonomics activities [10]. It can be hard to break this convenient design culture as subcontractor is employing software engineers in tight timeframe. In this context, we try a new incremental process to deal with early front-end development requirements.
To ensure usable and useful software, we empirically plan 3 user test increments with a minimum of 5 participants per end-user test [6]. These tests are based on the walkthrough ergonomics methodology [7] by using fictive operational scenarios. Scenario’s increments are built from simple and non-risky interactions (increment 1) to complex widget and data visualization interactions (increment 2 and 3). Scenario coverage (e.g. “1 + 2 + 3”) between increments allows us to compare if previous front-end modifications bring better ergonomics satisfaction through usability metrics. At the end of each increment, user observations are discussed with the project manager before integrating them to subcontractor’s agile development: considering usability gain over development planning, feasibility and effort.
3 Study Case: PRISM
PRISM is a web-based application inside the SCC eco-system. It basically executes procedure (list of command and automatic checks edited from SCOPE) to the real satellite or to a test platform such as avionic test bench, simulator, and other electrical ground system through the communication module (CMCS) (Fig. 2).
Thales Alenia Space has specified a graphical charter for every web applications and lot of reusable widgets have already been designed for other solutions (such as button, table, list, interactive menu, and so on). Therefore we decided to directly work with high fidelity mockups rather than having a first low fidelity prototype’s iteration. Prototypes were made using AXURE software. This tool creates interactive interfaces that can simulate expected scenario’s solution behavior and be used in our user testing sessions.
4 Process Inspection
4.1 User Testing Protocol
To investigate process relevance, it is important to set up a testing protocol. It allows us to compare data between increments. At the moment, two end-user test iterations have been performed (the third and last iteration will be done before product release). 10 distinct end-users participate to this analysis: 5 at first iteration and 7 at the second one (2 persons from increment 1 were involved as well in increment 2). Participants are all PRISM final end-users and come from different departments with different level of expertise. Each individual test session lasts 1 h and a half and follows the same walkthrough methodology [7]:
-
1.
Welcome participant, Informed consent form signature and present user centered approach
-
2.
Scenario reading and mockup limitation explanations (example: few auto scroll interactions could not be simulated)
-
3.
Participant interaction with the prototype performing scenario goals without any help from the ergonomist. If the user seems stuck after many tries, the ergonomist gives the solution but then the interaction is tagged as failed (0% or 100% in Figs. 7 and 8).
-
4.
System usability scale (SUS) survey to measure global software satisfaction
-
5.
Task satisfaction scale form and user remark debriefing
-
6.
Open questions and participant acknowledgments
4.2 User Satisfaction
System Usability Scale
At the end of each test, participants fill SUS form. SUS gives us the global satisfaction of the prototype. Rather than “Think aloud protocol”, participants were not allowed to give any observation before this step. This protocol aims to avoid any ergonomist’s explanation bias from participant satisfaction opinion (Figs. 5 and 6).
From increment 1 and 2, we can observe a big satisfaction deterioration: losing 17 points on SUS average score. The main empirical hypothesis raised from this drop is the higher complexity and the higher number of widgets provided between increments (see Figs. 3 and 4). For the next iteration, our SUS goal is to reach a score higher to 72/100 by moving secondary widgets out of first end-user’s eye exploration (and placing them into visible/hidden panel).
Task Satisfaction Scale
SUS doesn’t extract which technical task lower the satisfaction. So we decided to add a dirty homemade survey (rating from 1 to 5). This survey aims to prioritize which part of the software front-end need to be improved for the next iteration. A proper survey wasn’t investigated as it would have taken too much time in the test protocol.
From increment 1 and 2, five equivalent tasks were covered (e.g. “ouverture d’une procedure”) and four tasks were totally new. We can notice that modifications made between increments did bring better satisfaction for three of them (while the others stayed stable). By contrast with SUS score, it gives us a good feedback for our dirty UCD approach relevance. This chart will also be used to compare this analysis to iteration 3 and check task satisfaction progress.
4.3 Interaction Quality
Interaction quality is inspected through two usability metrics:
-
Effectiveness: does your participant succeeded or not to complete the interaction? (column “goal reached”)
-
Efficiency: does your participant make errors before completing his goal? And how long does he take to finish it? (column “error rate” and “time”)
The tables above list every interaction needed to complete the scenario at increment 1 and 2. The table colorization shows which operation need to be improved or modify in priority for the next increment. For instance, between mockup 1 and 2, a big front-end modification has been designed to open and initialize a procedure: passing from native browser “file explorer” to “custom procedure explorer” (interactions represented by #1.02 to #1.04 of Fig. 7 and #1.02 to #1.09 of Fig. 8). Surprisingly even if task satisfaction grew significantly (Fig. 9), the newer interface didn’t lower interaction error rate: 1, 6 at iteration 1 and 3, 5 at iteration.
On other hand, time measures didn’t bring any comparative added-value on interaction quality. Overall time evaluation must be changed to task-oriented time evaluation. For the next iteration, time should be sampled by task (column “scenario part”) and task should be kept consistent on next iterations. Also by using keystroke-level ergonomics method [9], we could analyze and compare the interaction performance to what an expert would have done.
4.4 Solution Convergence
Solution convergence is analyzed from the user remark number at each end of increment (see activity 4 Fig. 1).
User remark list sample. Each remark is tagged with the number of people who ask for it, from which department the participant comes from, when it has been identify (increment number), what is the empirical perceived ergonomics gain if integrated and what is the perceived complexity to implement it.
At the first increment, we report 97 user remarks on the tested prototype. At the second iteration, we count only 51 observations (approximately −50%) for a prototype which cover more complexity and tasks. This reduction shows a first solution convergence to the final user requirements. This preliminary analysis will be corroborated with the next increment results.
4.5 Continuous Improvement
Another important point measured in our process is the continuous improvement of PRISM interactions at each increments (Figs. 11 and 12).
From remark listing (Fig. 10), every user feedback is labeled by his integration status (according to interaction designer and project manager discussion): Accepted (green) meaning “sent directly to development requirements” - To be specified (blue) meaning “related to other software system specifications” - Delayed (purple) meaning “to be investigated at next iteration” - To be discussed (orange) meaning “to be discussed with the chef architect for feasibility investigation” - Rejected (red) meaning “remark is not relevant or in contradiction with previous taken decision”. From the two increment statements, we can deduce that the closer the project gets to the deadline, the fewer remarks are integrated into development (−54% accepted, +58% rejected and +80% to be discussed).
5 Conclusion
This paper presents the ongoing work of an early incremental user testing design approach. This dirty user centered design process allows front-end development at day one (as the first increment was prototyped and tested within 10 days). To assess this methodology relevance, we base our preliminary analysis on utility, usability and traceability metrics from two incremental user testing iterations.
As more complex features are integrated at each iteration, this approach doesn’t seem relevant to keep SUS score consistently high. Also from interaction efficiency and effectiveness metrics, no significant improvements were noticed. These results are mainly due to the incremental approach where designer has to focus on next increment requirements rather than improving previous ones. From user remark traceability statement, this process doesn’t as well give the opportunity to make major modifications as the development team gets closer to the release deadline.
However, homemade task satisfaction progress and user remark number analysis give a good feedback on the convergence of user needs. Furthermore, by integrating user in prototype testing loops, we identify one “game changer” idea which will be developed and test at iteration 3.
If development-first approach is requested and mandatory, this incremental user testing design approach can be applied as a dirty UCD process. Despite the lack of user research activities, our return on experience shows solution convergence and great feedback from participants but low capabilities on continuous improvement (interaction usability and remark integration) and some risks of user satisfaction drops.
References
Ergonomics of human-system interaction – Part 210: Human-centred design for interactive systems, ISO 9241-210 (2010). https://www.iso.org/standard/52075.html
Caplan, S.: Using focus group methodology for ergonomic design. Ergonomics 33(5), 527–533 (1990)
Muller, M.J., Kuhn, S.: Participatory design. Commun. ACM 36(6), 24–28 (1993)
Martinie, C., Palanque, P., Barboni, E., Ragosta, M.: Task-model based assessment of automation levels: application to space ground segments. In: 2011 IEEE International Conference on Systems, Man, and Cybernetics (SMC), pp. 3267–3273. IEEE, October 2011
Bornet, C., Brangier, E.: La méthode des personas: principes, intérêts et limites. Bulletin de psychologie (2), 115–134 (2013)
Nielsen, J.: Why you only need to test with 5 users (2000)
Wharton, C.: The cognitive walkthrough method: a practitioner’s guide. In: Usability Inspection Methods (1994)
Van Someren, M.W., Barnard, Y.F., Sandberg, J.A.C.: The think aloud method: a practical approach to modeling cognitive (1994)
Card, S.K., Moran, T.P., Newell, A.: The keystroke-level model for user performance time with interactive systems. Commun. ACM 23(7), 396–410 (1980)
Holtzblatt, K., Beringer, J., Baker, L.: Rapid user centered design techniques: challenges and solutions. In: CHI 2005 Extended Abstracts on Human Factors in Computing Systems, pp. 2037–2038. ACM, April 2005
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 IFIP International Federation for Information Processing
About this paper
Cite this paper
Tolle, J. (2019). Early Incremental User Testing Design Approach Validation for Satellite Command Center’s Application. In: Bogdan, C., Kuusinen, K., Lárusdóttir, M., Palanque, P., Winckler, M. (eds) Human-Centered Software Engineering. HCSE 2018. Lecture Notes in Computer Science(), vol 11262. Springer, Cham. https://doi.org/10.1007/978-3-030-05909-5_19
Download citation
DOI: https://doi.org/10.1007/978-3-030-05909-5_19
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-05908-8
Online ISBN: 978-3-030-05909-5
eBook Packages: Computer ScienceComputer Science (R0)