Keywords

1 Introduction

User centered design is a well-known software conception standard in ergonomics and human factors fields (ISO 9241-210) [1]. To guarantee system’s utility and usability, it implies four stages where end-user is integrated within many ergonomics methods such as focus group [2], participatory design [3], interviews, job task-model based analysis [4], persona specification [5] and so on. In some constrained industrial context, deploying every practices cannot be possible. Thales Alenia Space believes that every process can be tailored and scaled to fit every project. For our new satellite command center (SCC) procedure executor named PRISM, we try a dirty UCD approach based on incremental mockups and user testing only. As user requirements were taken from our previous product version, this process enables earlier front-end development than classical ergonomics methodology.

This testimonial explains first how we integrated UCD activities within agile development process. Then we briefly present our study case PRISM and its incremental mockups. Next, we inspect our process relevance from usability metrics and user feedbacks. Does this approach lead to solution convergence? Does it bring high quality interaction and high user satisfaction? Does it allow continuous interaction improvement? From these problematic, the paper concludes on a preliminary analysis of this dirty UCD approach (ongoing work). Please note this article doesn’t list any SCC user needs. It only aims to analyze the process relevance from ergonomics point of view.

2 Process Overview

Industry often works with subcontractors while developing products. When interaction designers are not involved in project planning phase, it may results to a development-first approach without much consideration for ergonomics activities [10]. It can be hard to break this convenient design culture as subcontractor is employing software engineers in tight timeframe. In this context, we try a new incremental process to deal with early front-end development requirements.

To ensure usable and useful software, we empirically plan 3 user test increments with a minimum of 5 participants per end-user test [6]. These tests are based on the walkthrough ergonomics methodology [7] by using fictive operational scenarios. Scenario’s increments are built from simple and non-risky interactions (increment 1) to complex widget and data visualization interactions (increment 2 and 3). Scenario coverage (e.g. “1 + 2 + 3”) between increments allows us to compare if previous front-end modifications bring better ergonomics satisfaction through usability metrics. At the end of each increment, user observations are discussed with the project manager before integrating them to subcontractor’s agile development: considering usability gain over development planning, feasibility and effort.

3 Study Case: PRISM

PRISM is a web-based application inside the SCC eco-system. It basically executes procedure (list of command and automatic checks edited from SCOPE) to the real satellite or to a test platform such as avionic test bench, simulator, and other electrical ground system through the communication module (CMCS) (Fig. 2).

Thales Alenia Space has specified a graphical charter for every web applications and lot of reusable widgets have already been designed for other solutions (such as button, table, list, interactive menu, and so on). Therefore we decided to directly work with high fidelity mockups rather than having a first low fidelity prototype’s iteration. Prototypes were made using AXURE software. This tool creates interactive interfaces that can simulate expected scenario’s solution behavior and be used in our user testing sessions.

4 Process Inspection

4.1 User Testing Protocol

To investigate process relevance, it is important to set up a testing protocol. It allows us to compare data between increments. At the moment, two end-user test iterations have been performed (the third and last iteration will be done before product release). 10 distinct end-users participate to this analysis: 5 at first iteration and 7 at the second one (2 persons from increment 1 were involved as well in increment 2). Participants are all PRISM final end-users and come from different departments with different level of expertise. Each individual test session lasts 1 h and a half and follows the same walkthrough methodology [7]:

  1. 1.

    Welcome participant, Informed consent form signature and present user centered approach

  2. 2.

    Scenario reading and mockup limitation explanations (example: few auto scroll interactions could not be simulated)

  3. 3.

    Participant interaction with the prototype performing scenario goals without any help from the ergonomist. If the user seems stuck after many tries, the ergonomist gives the solution but then the interaction is tagged as failed (0% or 100% in Figs. 7 and 8).

  4. 4.

    System usability scale (SUS) survey to measure global software satisfaction

  5. 5.

    Task satisfaction scale form and user remark debriefing

  6. 6.

    Open questions and participant acknowledgments

4.2 User Satisfaction

System Usability Scale

At the end of each test, participants fill SUS form. SUS gives us the global satisfaction of the prototype. Rather than “Think aloud protocol”, participants were not allowed to give any observation before this step. This protocol aims to avoid any ergonomist’s explanation bias from participant satisfaction opinion (Figs. 5 and 6).

From increment 1 and 2, we can observe a big satisfaction deterioration: losing 17 points on SUS average score. The main empirical hypothesis raised from this drop is the higher complexity and the higher number of widgets provided between increments (see Figs. 3 and 4). For the next iteration, our SUS goal is to reach a score higher to 72/100 by moving secondary widgets out of first end-user’s eye exploration (and placing them into visible/hidden panel).

Task Satisfaction Scale

SUS doesn’t extract which technical task lower the satisfaction. So we decided to add a dirty homemade survey (rating from 1 to 5). This survey aims to prioritize which part of the software front-end need to be improved for the next iteration. A proper survey wasn’t investigated as it would have taken too much time in the test protocol.

From increment 1 and 2, five equivalent tasks were covered (e.g. “ouverture d’une procedure”) and four tasks were totally new. We can notice that modifications made between increments did bring better satisfaction for three of them (while the others stayed stable). By contrast with SUS score, it gives us a good feedback for our dirty UCD approach relevance. This chart will also be used to compare this analysis to iteration 3 and check task satisfaction progress.

4.3 Interaction Quality

Interaction quality is inspected through two usability metrics:

  • Effectiveness: does your participant succeeded or not to complete the interaction? (column “goal reached”)

  • Efficiency: does your participant make errors before completing his goal? And how long does he take to finish it? (column “error rate” and “time”)

The tables above list every interaction needed to complete the scenario at increment 1 and 2. The table colorization shows which operation need to be improved or modify in priority for the next increment. For instance, between mockup 1 and 2, a big front-end modification has been designed to open and initialize a procedure: passing from native browser “file explorer” to “custom procedure explorer” (interactions represented by #1.02 to #1.04 of Fig. 7 and #1.02 to #1.09 of Fig. 8). Surprisingly even if task satisfaction grew significantly (Fig. 9), the newer interface didn’t lower interaction error rate: 1, 6 at iteration 1 and 3, 5 at iteration.

On other hand, time measures didn’t bring any comparative added-value on interaction quality. Overall time evaluation must be changed to task-oriented time evaluation. For the next iteration, time should be sampled by task (column “scenario part”) and task should be kept consistent on next iterations. Also by using keystroke-level ergonomics method [9], we could analyze and compare the interaction performance to what an expert would have done.

4.4 Solution Convergence

Solution convergence is analyzed from the user remark number at each end of increment (see activity 4 Fig. 1).

Fig. 1.
figure 1

‘Early incremental user testing design approach’ activities

Fig. 2.
figure 2

PRISM simplified eco-system

Fig. 3.
figure 3

Increment 1: main screen which displays direct feedback of the procedure execution

Fig. 4.
figure 4

Increment 2: main screen which displays direct feedback of the procedure execution and provides complementary services to end-user

Fig. 5.
figure 5

French translated SUS result from the 5 participants of increment 1.

Fig. 6.
figure 6

French translated SUS result from the 7 participants of increment 2. Note that one participant results have been removed from average score due to inconsistent answers.

Fig. 7.
figure 7

Bar chart high-level functionality satisfaction result from both iterations (1 & 2). From an empiric decision, every score below 3, 4/5 is marked “has to be improved”

Fig. 8.
figure 8

Efficiency and effectiveness report from iteration 1 (5 participants)

Fig. 9.
figure 9

Efficiency and effectiveness report from iteration 2 (7 participants)

Fig. 10.
figure 10

User remark list sample. Each remark is tagged with the number of people who ask for it, from which department the participant comes from, when it has been identify (increment number), what is the empirical perceived ergonomics gain if integrated and what is the perceived complexity to implement it.

At the first increment, we report 97 user remarks on the tested prototype. At the second iteration, we count only 51 observations (approximately −50%) for a prototype which cover more complexity and tasks. This reduction shows a first solution convergence to the final user requirements. This preliminary analysis will be corroborated with the next increment results.

4.5 Continuous Improvement

Another important point measured in our process is the continuous improvement of PRISM interactions at each increments (Figs. 11 and 12).

Fig. 11.
figure 11

Remark traceability chart iteration 1 (Color figure online)

Fig. 12.
figure 12

Remark traceability chart iteration 2 (Color figure online)

From remark listing (Fig. 10), every user feedback is labeled by his integration status (according to interaction designer and project manager discussion): Accepted (green) meaning “sent directly to development requirements” - To be specified (blue) meaning “related to other software system specifications” - Delayed (purple) meaning “to be investigated at next iteration” - To be discussed (orange) meaning “to be discussed with the chef architect for feasibility investigation” - Rejected (red) meaning “remark is not relevant or in contradiction with previous taken decision”. From the two increment statements, we can deduce that the closer the project gets to the deadline, the fewer remarks are integrated into development (−54% accepted, +58% rejected and +80% to be discussed).

5 Conclusion

This paper presents the ongoing work of an early incremental user testing design approach. This dirty user centered design process allows front-end development at day one (as the first increment was prototyped and tested within 10 days). To assess this methodology relevance, we base our preliminary analysis on utility, usability and traceability metrics from two incremental user testing iterations.

As more complex features are integrated at each iteration, this approach doesn’t seem relevant to keep SUS score consistently high. Also from interaction efficiency and effectiveness metrics, no significant improvements were noticed. These results are mainly due to the incremental approach where designer has to focus on next increment requirements rather than improving previous ones. From user remark traceability statement, this process doesn’t as well give the opportunity to make major modifications as the development team gets closer to the release deadline.

However, homemade task satisfaction progress and user remark number analysis give a good feedback on the convergence of user needs. Furthermore, by integrating user in prototype testing loops, we identify one “game changer” idea which will be developed and test at iteration 3.

If development-first approach is requested and mandatory, this incremental user testing design approach can be applied as a dirty UCD process. Despite the lack of user research activities, our return on experience shows solution convergence and great feedback from participants but low capabilities on continuous improvement (interaction usability and remark integration) and some risks of user satisfaction drops.