Abstract
Electroencephalography based brain computer interface has enabled communication of human’s intention to a computer directly from the brain by decoding signatures that relay the intention information. Error-related potential has been adopted as a signature for natural communication and performance monitoring, among others. In this work, we investigate the use of error-related potential as an input channel to transfer human preference of a strategical advisory to a computer. Air traffic control task was used as a case study to make an empirical inquiry of error-related potential for higher level cognitive tasks (i.e. situation awareness in air traffic control tasks). The experimental task requires the subjects to monitor and assess air traffic scenarios presented on a simulated environment that provides conflict resolution advisories. The task is known to be highly mental demanding as it requires continuous situation awareness of the traffic. An interface and experimental protocol were developed for this experiment to validate that error-related potential can be used a new channel for preference. The implementation of the complete design was described together with the experimental evidence of error-related potential. According to the results, we found error-related potential that is in line with existing literature. We also discussed how the preliminary findings of this work can be used as an integral part of an intelligent conflict resolution advisory system that can learn from human preference and duplicate the decision making.
This research has been partially supported under Air Traffic Management Research Institute (NTU-CAAS) Grant No. M4062429.052.
You have full access to this open access chapter, Download conference paper PDF
Similar content being viewed by others
Keywords
- Brain computer interface
- Error-related potential
- Decision network
- Air traffic management
- Air traffic control
- Conflict detection and resolution
1 Introduction
Electroencephalography (EEG), a widely used brain-computer interface (BCI) that captures electrical brain activities, allows communication of human’s intention directly from the human brain to computer system. To date, a number of brain patterns (e.g. P300 [21], motor imagery [20], error-related potential (ErrP) [17]) have been identified and utilized for BCI applications. Among these patterns, ErrP is naturally elicited when the brain observes an erroneous or unexpected behavior without explicit training or instructing participant to generate it [28]. Hence, ErrP has the advantage of facilitating natural and intuitive interaction between the brain and computer. A few pioneering work has employed ErrP in a human-machine setting. In [24], ErrP was monitored and decoded to correct robot mistakes. ErrP was also used to map human gesture to robot’s action in an interactive setting [14]. Moreover, [28] used ErrP to select the heading of the vehicle in real-world driving task.
Existing BCI literature has been demonstrating the use of BCI to communicate instantaneous/operational command (i.e. left/right, start/stop) to control an external system on various tasks. Nonetheless, operational level BCI control is laborious, time-consuming and mentally demanding. Besides, even instances of similar tasks require human to generate a similar sequence of BCI commands repeatedly. However, there is little research work extends BCI to communicate tactical or strategical level commands, which can perform a set of sequential actions and generalize to solve similar tasks. Hence, we attempt to investigate the possibility of creating set of higher level instruction, decision network or preference using ErrP based BCI.
In air traffic management literature, BCI have been applied in air traffic management to monitor air traffic controller’s (ATCO) performance. EEG was utilized to derive objective bio-marker of workload [2,3,4]. The findings are useful to ensure the situation awareness of ATCOs. Detector can then be developed to alert any sign of sleepiness or loss of vigilance. fNIRS was adopted to measure maturity and expertise of air traffic controllers [5] which will aid the training and selection process. While those work use BCI in a passive manner, our work has a very different objective which aims to use BCI to transfer preference actively.
In this work, we select one the most challenging task in air traffic control (i.e. conflict detection and resolution) to be an ErrP case study of higher level cognitive task. This task requires air traffic controller (ATCO) to maintain situation awareness of current and future air traffic condition to ensure a safe and efficient flow of every aircraft in a shared air space. Many researches and developments have been performed to develop assistance tools for ATCOs to reduce their workload whilst improving their performance. The pioneering researches relied on mathematical models of air-crafts, conflict scenarios, and airspace structure to compute conflict resolution strategy. An extensive model based approaches can be found in [15]. Recently, second-order cone programming [26], space-time Prism [9] approach, model predictive control (MPC) [13, 27], surrounding traffic analysis [22], and large scale conflict resolution models for velocity maneuver [1] and 3D conflict resolution [16]. These mathematical models can hardly scale up for a big amount of aircraft and might fail to describe the complete dynamics of air traffic. Moreover, these automated tools were not fully trusted as most models behave like a black-box to ATCOs. Besides, the advisories might be very different from ATCO’s expectation that leads to their low acceptance rate [11]. Hence, ATCOs have to remain taking active control in the management of air traffic.
While mathematical models showed their limitation in incorporating human preferences or strategies in their solution, artificial intelligence (AI) (e.g. deep learning and reinforcement learning (RL)) has achieved superhuman level in variety of strategical tasks (e.g. diagnosing a number of cancers, playing the game of Go and Atari games, etc.). Recently, the literature in behavior cloning [12] and inverse RL [8, 18, 19, 23] have demonstrated machine abilities in mimicking the expert’s behaviors from the demonstrations or even infer the the reward function of human strategies [7, 10, 25]. Following this line of research, RL can be adopted in air traffic control task to learn how ATCOs perform the conflict resolution.
In this study, we aims to integrate the advancement of BCI and RL. Our goal is to develop an BCI framework where a human can communicate and construct a goal-oriented sequential decision-making or preference command using ErrP signature. RL can be used as an engine to incorporate human’s preference in learning model for conflict resolution. In this paper, we limit our work to fill the research gap on the empirical inquiry of error-related potential for higher level cognitive tasks (i.e. situation awareness in air traffic control tasks). In order to investigate our hypothesis that ErrP can be adopted in this air traffic monitoring task, we developed a simulated air traffic environment which can simulate different configuration of conflicts and visualize projected trajectories of traffic scenarios as well as advisory trajectory. The environment allows subject to monitor and assess the advisory naturally. We also design experimental protocol that facilitates the generation of ErrP that encapsulate ATCO’s preference. The mapping between the ErrP and preference as well as how they can be useful will be discussed in succeeding parts of this paper.
The main contribution of this work includes (1) the study of ErrP for a higher level planning task (i.e. air traffic monitoring tasks) compared to previous work on instantaneous controlling tasks; (2) the development of simulated air traffic environment to emulate real-world traffic (3) the design of visual interfaces and experimental protocol for subjects to input their preferences using ErrP; (4) experimental evidence that our proposed protocol triggered brain signature that is consistent with existing ErrP literature.
The rest of this paper is structured as follows. Section 2 describes the methodology and how the whole BCI framework is designed and implemented to achieve the experimental goal of this work. It is followed by Sect. 3 that shows the experimental results of the ErrP generated by the subject. Finally, Sect. 4 discusses how the findings and experimental evidence can be used to adapt advisory tool.
2 Methodology
This is an exploratory study to investigate the extension of ErrP for higher level cognitive task. We implemented the BCI framework and design experimental protocol to validate the applicability of ErrP for a real-world problem in air traffic management. The experimental setup and paradigm are illustrated in Fig. 1.
2.1 Simulated Environment and Scenario Generation
The simulated environment configures conflict scenarios with two air-craft as shown in Fig. 2. For simplicity, the airspace was assumed to a circular area. In the airspace presented to the subject, there is an ownship and an intruder together with their projected trajectories. We restrict the two air-crafts to fly as the same lateral speed in the environment. The conflict configuration can be characterized by the conflict angle and closest point of approach. Advisory tool can propose a heading change maneuver to resolve the conflict. An example is shown in Fig. 2.
2.2 Experimental Protocol
The subject was seated in front of a screen displaying a variety of air traffic scenarios. The time-line of the experiment can be found in Fig. 3. A trial starts with a little dot in the middle of the screen with a audio cue. The color of the dot is hint of the quality of the solution showing next (green indicates good advisory while red indicates bad advisory). Subsequently, an air traffic scenario with resolution advisory is presented to the subject. The ownship is presented using black line while the intruder using gray line. On the ownship trajectory, there are 3 little circles. Subject monitor and assess the advisory trajectory of the ownship for 5 s. Next, the middle circle changes from unfilled to filled circle to cue the subject to get ready for the next visual event. The middle filled circle might either make a positional change in the same direction as the advisory trajectory or the opposite direction. Subject was told that the same direction signifies acceptance of the advisory and the opposite direction indicate rejection of the advisory. However, the direction is assigned randomly by design. We assume ErrP to be triggered when the assigned direction does not match with what the subject expects.
2.3 EEG Acquisition and Processing
When the subject was performing the task, EEG was acquired with a sampling frequency of 250 Hz using BrainAmp MR plus EEG device. We custom-made a circuit for EEG event marker using NI USB-6001. It facilitates our simulated environment to send triggers to the Brain Vision Recorder. In this study, we select and analyze the channel C2. Notch filter was applied to remove the line noise. The signal was then band-pass filtered between 1 and 5 Hz. During the experiment, subject was instructed to prevent muscle movement that can induce artifact to the signal. The recorded signal is subsequently post-check to reject artifact trials. EEGLAB was used to extract the signal segments 0.5 s before and 1.5 s after the positional change on of the circle on the ownship trajectory.
3 Results
The event related potential of the experiment was shown in Fig. 4. Trials with positional changes that match subject’s expectation are shown in Fig. 4 (left) and trials with positional changes that do not match subject’s expectation are shown in Fig. 4 (right). By comparing the two figures, consistent event related potentials were observed for trials with positional changes that do not match subject’s expectation. The difference between ErrP and Non ErrP was shown in Fig. 5. The pattern of the potential is similar to the ErrP reported in [6], a negative deflection followed by positive peak. Besides, we also observed longer ErrP latency compared to [6].
4 Discussion
The result of the experiment demonstrated that ErrP can be extended to conflict resolution in air traffic control, a higher level cognitive task. Based on the ErrP finding, we obtained experimental evidence that human preference can be encapsulated into the ErrP signature. Decoding of this signature can be used to adapt an intelligent agent to behave in the way human will trust. While the resolution advisory was hand designed in this experiment, it is possible implement the advisory model using inverse RL, where it can be proposing random maneuver initially and gaining air traffic control skill by iterative interaction with ATCOs.
In the succeeding parts of the paper, we will discuss how the ErrP obtained from our protocol can be used to adapt the advisory tool.
4.1 Mapping Between ErrP and Preference
As ErrP is supposed to be triggered when the outcome of the positional change does not match the expectation. Hence, the expected positional change can be derived using the existence of ErrP and the outcome of the assigned positional change. The mapping between ErrP and preference (accept or reject) can be found in the Table 1.
4.2 ErrP as Building Block for Reinforcement Learning
The mapping obtained from Subsect. 4.1 can be used as a reinforcement signal to adapt advisory tool. One of the straightforward approach is to adopt reinforcement learning and define the reward function of reinforcement learning using ATCO’s acceptability.
Let define the reward function \(\mathcal {R}_0\) as follows:
The reward function \(\mathcal {R}_0\) can be used in addition to the environmental reward function \(\mathcal {R}_1\) which access the quality of the resolution based on common criteria such as deviation, travel distance, travel time or fuel consumption. These rewards can be combined by weighted sum.
4.3 Possible Future Work
As the focus of this work is to study ErrP for higher level cognitive task, several restriction was made to control and simplify the simulated environment. Relaxing the restriction and assumption part by part is important to improve the practicality of this work for solving real-world air traffic problem.
While the framework is designed for air traffic task, it is also applicable to amyotrophic lateral sclerosis (ALS) patients where they can construct their goal-oriented sequential decision making command.
References
Allignol, C., Barnier, N., Durand, N., Gondran, A., Wang, R.: Large scale 3D en-route conflict resolution. In: ATM Seminar, 12th USA/Europe Air Traffic Management R&D Seminar
Aricò, P., et al.: Adaptive automation triggered by eeg-based mental workload index: a passive brain-computer interface application in realistic air traffic control environment. Front. Hum. Neurosci. 10, 539 (2016)
Ayaz, H., et al.: Cognitive workload assessment of air traffic controllers using optical brain imaging sensors. In: Advances in Understanding Human Performance: Neuroergonomics, Human Factors Design, and Special Populations, pp. 21–31 (2010)
Brookings, J.B., Wilson, G.F., Swain, C.R.: Psychophysiological responses to changes in workload during simulated air traffic control. Biol. Psychol. 42(3), 361–377 (1996)
Bunce, S.C., et al.: Implementation of fNIRS for monitoring levels of expertise and mental workload. In: Schmorrow, D.D., Fidopiastis, C.M. (eds.) FAC 2011. LNCS (LNAI), vol. 6780, pp. 13–22. Springer, Heidelberg (2011). https://doi.org/10.1007/978-3-642-21852-1_2
Chavarriaga, R., Ferrez, P.W., Millán, J.R.: To Err is human: learning from error potentials in brain-computer interfaces. In: Wang, R., Shen, E., Gu, F. (eds.) Advances in Cognitive Neurodynamics ICCN 2007. Springer, Dordrecht (2008). https://doi.org/10.1007/978-1-4020-8387-7_134
Gao, Y., et al.: Reinforcement learning from imperfect demonstrations. arXiv preprint arXiv:1802.05313 (2018)
Hadfield-Menell, D., Russell, S.J., Abbeel, P., Dragan, A.: Cooperative inverse reinforcement learning. In: Advances in Neural Information Processing Systems, pp. 3909–3917 (2016)
Hao, S., Cheng, S., Zhang, Y.: A multi-aircraft conflict detection and resolution method for 4-dimensional trajectory-based operation. Chinese Journal of Aeronautics 31(7), 1579–1593 (2018). http://www.sciencedirect.com/science/article/pii/S1000936118301705
Hester, T., et al.: Deep Q-learning from demonstrations. In: Thirty-Second AAAI Conference on Artificial Intelligence (2018)
Hilburn, B., Westin, C., Borst, C.: Will controllers accept a machine that thinks like they think? the role of strategic conformance in decision aiding automation. Air Traffic Control Q. 22(2), 115–136 (2014)
Ho, J., Ermon, S.: Generative adversarial imitation learning. In: Advances in Neural Information Processing Systems, pp. 4565–4573 (2016)
Jilkov, V.P., Ledet, J.H., Li, X.R.: Multiple model method for aircraft conflict detection and resolution in intent and weather uncertainty. IEEE Transact. Aerosp. Electron. Syst. 55, 1–1 (2018)
Kim, S.K., Kirchner, E.A., Stefes, A., Kirchner, F.: Intrinsic interactive reinforcement learning-using error-related potentials for real world human-robot interaction. Sci. Rep. 7(1), 17562 (2017)
Kuchar, J.K., Yang, L.C.: A review of conflict detection and resolution modeling methods. IEEE Transact. Intell. Transp. Syst. 1(4), 179–189 (2000)
Liu, Z., Cai, K., Zhu, X., Tang, Y.: Large scale aircraft conflict resolution based on location network. In: 2017 IEEE/AIAA 36th Digital Avionics Systems Conference (DASC), pp. 1–8 (2017)
Nieuwenhuis, S., Ridderinkhof, K.R., Blom, J., Band, G.P., Kok, A.: Error-related brain potentials are differentially related to awareness of response errors: evidence from an antisaccade task. Psychophysiology 38(5), 752–760 (2001)
Odom, P., Natarajan, S.: Active advice seeking for inverse reinforcement learning. In: Proceedings of the 2016 International Conference on Autonomous Agents and Multiagent Systems, pp. 512–520. International Foundation for Autonomous Agents and Multiagent Systems (2016)
Pan, X., Shen, Y.: Human-interactive subgoal supervision for efficient inverse reinforcement learning. In: Proceedings of the 17th International Conference on Autonomous Agents and MultiAgent Systems, pp. 1380–1387. International Foundation for Autonomous Agents and Multiagent Systems (2018)
Pfurtscheller, G., Neuper, C.: Motor imagery and direct brain-computer communication. Proc. IEEE 89(7), 1123–1134 (2001)
Polich, J.: Updating P300: an integrative theory of P3a and P3b. Clin. Neurophysiol. 118(10), 2128–2148 (2007)
Radanovic, M., Eroles, M.A.P., Koca, T., Gonzalez, J.J.R.: Surrounding traffic complexity analysis for efficient and stable conflict resolution. Transp. Res. Part C Emerg. Technol. 95, 105–124 (2018). http://www.sciencedirect.com/science/article/pii/S0968090X18302353
Ross, S., Gordon, G., Bagnell, D.: A reduction of imitation learning and structured prediction to no-regret online learning. In: Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics, pp. 627–635 (2011)
Salazar-Gomez, A.F., DelPreto, J., Gil, S., Guenther, F.H., Rus, D.: Correcting robot mistakes in real time using EEG signals. In: 2017 IEEE International Conference on Robotics and Automation (ICRA), pp. 6570–6577. IEEE (2017)
Schaal, S.: Learning from demonstration. In: Advances in Neural Information Processing Systems, pp. 1040–1046 (1997)
Yang, Y., Zhang, J., Cai, K., Prandini, M.: Multi-aircraft conflict detection and resolution based on probabilistic reach sets. IEEE Transact. Control Syst. Technol. 25(1), 309–316 (2017)
Yokoyama, N.: Decentralized conflict detection and resolution using intent-based probabilistic trajectory prediction. In: AIAA SciTech Forum, American Institute of Aeronautics and Astronautics (2018). https://doi.org/10.2514/6.2018-1857
Zhang, H., Chavarriaga, R., Khaliliardali, Z., Gheorghe, L., Iturrate, I., d R Millán, J.: EEG-based decoding of error-related brain activity in a real-world driving task. J. Neural Eng. 12(6), 066028 (2015)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this paper
Cite this paper
Goh, S.K., Tran, N.P., Pham, DT., Alam, S., Izzetoglu, K., Duong, V. (2019). Construction of Air Traffic Controller’s Decision Network Using Error-Related Potential. In: Schmorrow, D., Fidopiastis, C. (eds) Augmented Cognition. HCII 2019. Lecture Notes in Computer Science(), vol 11580. Springer, Cham. https://doi.org/10.1007/978-3-030-22419-6_27
Download citation
DOI: https://doi.org/10.1007/978-3-030-22419-6_27
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-22418-9
Online ISBN: 978-3-030-22419-6
eBook Packages: Computer ScienceComputer Science (R0)