Abstract
The paper presents a multi-agent system that learns to manage the re-sources of an unmanned spacecraft. Each agent controls a subsystem and learns to optimise its resources. The agents can co-ordinate their actions to satisfy user requests. Co-ordination is achieved by exchanging sched-uling information between agents. Resource management is implemented using two reinforcement learning techniques: the Montecarlo and the Q-learning. The paper demonstrates how the approach can be used to model the imaging system of a spacecraft. The environment is represented by agents which control the spacecraft sub-systems involved in the imaging activity. The agent in charge of the resource management senses the information regarding the resource requested, the resource conflicts and the resource availability. Scheduling of resources is learnt when all subsystems are fully functional and when resources are reduced by random failures.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
C.C. Cheng and S.F. Smith. “Applying constraint satisfaction techniques to job shop scheduling”. “CMU-RI-TR-95-03”, CMU, January 1995.
A. Chavez, A. Moukas, and P. Maes. “Challenger: A multi-agent system for distributed resource allocation”. In 1 st Conference on Autonomous Agents, pages 323–331. ACM Press, February 5–8 1997.
R.H. Crites and A.G. Barto. “Improving elevator performance using reinforcement learning”. In D.S. Touretzky, M.C. Mozer, and M.E. Hasselmo, editors, Advances in Neural Information Processing Systems, pages 1017–1023. MITPress, 1996.
D.E. Neiman and V.R. Lesser. “A Co-operative repair method for a distributed scheduling system”. In 3 rd International Conference on Artificial Planning Systems, pages 166–173. AAAI Press, 1996.
A.S. Fukunaga, G. Rabideau, and S. Chien. “SPEN: An Application Framework for Automated Planning and Scheduling of Spacecraft Control and Operations”. In i-SAIRAS, pages 181–187, Tokyo, Japan, 1997.
M. Fox. “ISIS, a retrospective”. In M. Zweben and M. Fox, editors, Intelligent scheduling, pages 3–28. Morgan Kaufmann Press, 1994.
M. Zweben, B. Daun, E. Davis, and M. Daele. “Scheduling and re-scheduling with iterative repair”. In M. Zweben and M. Fox, editors, Intelligent scheduling, pages 241–256. Morgan Kaufmann Press, 1994.
N. Sadeh. “Micro-opportunistic scheduling”. In M. Zweben and M. Fox, editors, Intelligent scheduling, pages 99–136. Morgan Kaufmann Press, 1994.
Y. Shoham. “Agent-Oriented Programming”. Artificial Intelligence, 60:51–92, 1993.
S.F. Smith. “Reactive scheduling system”. In D.E. Brown and W.T. Scherer, editors, Intelligent scheduling systems. Kluwer Press, 1995.
Richard S. Sutton and Andrew G. Barto. “Reinforcement Learning An Introduction”. MITPress, 1998.
W. Zhang and T.G. Dietterich. “Solving combinatorial optimisation tasks by reinforcement learning: A general methodology applied to resource constrained scheduling”. Artificial Intelligence Research, 1, 1998.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2000 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Monekosso, N., Remagnino, P. (2000). Autonomous Spacecraft Resource Management: A Multi-Agent Approach. In: Lamma, E., Mello, P. (eds) AI*IA 99: Advances in Artificial Intelligence. AI*IA 1999. Lecture Notes in Computer Science(), vol 1792. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-46238-4_26
Download citation
DOI: https://doi.org/10.1007/3-540-46238-4_26
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-67350-7
Online ISBN: 978-3-540-46238-5
eBook Packages: Springer Book Archive