A hybrid model-free approach for the near-optimal intrusion response control of non-stationary systems

Given the always increasing size of computer systems, manually protecting them in case of attacks is unfeasible and error-prone. For this reason, until now, several model-based Intrusion Response Systems (IRSs) have been proposed with the purpose of limiting the amount of work of the system administ...

Full description

Saved in:
Bibliographic Details
Published in:Future generation computer systems Vol. 109; pp. 111 - 124
Main Authors: Iannucci, Stefano, Cardellini, Valeria, Barba, Ovidiu Daniel, Banicescu, Ioana
Format: Journal Article
Language:English
Published: Elsevier B.V 01.08.2020
Subjects:
ISSN:0167-739X, 1872-7115
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Given the always increasing size of computer systems, manually protecting them in case of attacks is unfeasible and error-prone. For this reason, until now, several model-based Intrusion Response Systems (IRSs) have been proposed with the purpose of limiting the amount of work of the system administrators. However, since the most advanced IRSs adopt a stateful approach, they are subject to what Richard Bellman defined as the curse of dimensionality. Furthermore, modern computer systems are non-stationary, that is, they are subject to frequent changes in their configuration and in their software base, which in turn could make a model-based approach ineffective due to deviations in system behavior with respect to the model. In this paper we propose, to the best of our knowledge, the first approach based on deep reinforcement learning for the implementation of a hybrid model-free IRS. Experimental results show that the proposed IRS is able to deal with non-stationary systems, while reducing the time needed for the computation of the defense policies by orders of magnitude with respect to model-based approaches, and being still able to provide near-optimal rewards. •Approach based on transfer learning, aimed at decreasing the training time•Performance comparison of Q-Learning and Deep Q-Learning on non-stationary systems•Analysis of the initial tuning of the parameters of Deep Q-Learning algorithm
ISSN:0167-739X
1872-7115
DOI:10.1016/j.future.2020.03.018