- Student: Manuel Rickli
- Supervisor: Pontus Johnson
- Examiner: György Dán
- Active members:
- Status: Finished
- Timeline: VT-21
Attack simulation provides insight over an attacker's capabilities within a system.
In order to have a more meaningful simulation, the attacker should realistically only
be able to partially observe the system and make decisions based on the current
information available.
This work aims to clarify the question of how well a neural network can approximate the
optimal policy of a partially observable attack graph. The generated attack graphs used for
supervised learning are heavily inspired by coreLang.
- Generate attack graphs with prefixes
train
,val
andtest
and save them inAttackGraphs
- Modify the training configuration in
value_approximator/gat/train.py
- Run the training
- Optionally, test the prediction accuracy with the trained model that saved in
models/binaries
(the final model) ormodels/checkpoints
- Build attack graph from generated model
- Calculated value function
- Adapt GNN for attack graphs
- First GNN learning on attack graphs
- Analyse results
- Reiterate learning and conclude
This is a project run by the Software Systems Architecture and Security research group within the Division of Network and Systems Engineering at the Department of Computer Science at the School of Electrical Engineering and Computer Science @ KTH university.
For more of our projects, see the SSAS page at github.com.