Skip to content

TTCs as Reward #165

@sandorstormen

Description

@sandorstormen

One way you can represent cost for RL algorithms is to adjust the reward function. We should implement an option to represent TTCs in terms of reward. The reward function (or more aptly named utility function) should give the negative TTC when performing an attack step that always succeeds.

There are two formulations for this:

$r(a) = -\texttt{ttc value}$, where $\texttt{ttc value}\sim\text{TTCDist}(a)$

$r(a) = -\texttt{ttc value}$, where $\texttt{ttc value} = \mathbb{E}[\text{TTCDist}(a)]$

such that $r(\cdot)$ is the reward/utility function and $a$ is an attack step.

Let me note here that this has already been done by Sandor and Manuel with a simplified coreLang.

Metadata

Metadata

Assignees

Labels

enhancementNew feature or request

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions