Comments (2)
It seems that when the Gym environment is initialized the following is passed as a winning reward:
winning_reward=5000.0
(line 396 in cyberbattle_env.py). If the agent reaches the goal this is added as a reward overriding the value of the node. This does not happen when you play the interactive game, which returns the string "FLAG: flag discovered!" and just adds the node value which is 1000
.
Can someone please confirm that this explanation is correct and it is a matter of just adding 4000 to the final reward to compare the results?
from cyberbattlesim.
@MariaRigaki The explanation is correct. The attacker gets the final winning reward if one is specified when the environment is created. The reason why this does not happen in the interactive game is because in notebook_benchmark-chain.ipynb
the environment that gets instantiated is CyberBattleToyCtf-v0
. This environment is defined in __init__.py
as:
register(
id='CyberBattleToyCtf-v0',
cyberbattle_env_identifiers=toy_ctf.ENV_IDENTIFIERS,
entry_point='cyberbattle._env.cyberbattle_toyctf:CyberBattleToyCtf',
kwargs={'defender_agent': None,
'attacker_goal': AttackerGoal(own_atleast=6),
'defender_goal': DefenderGoal(eviction=True)
},
)
which does not have a winning_reward
parameter.
from cyberbattlesim.
Related Issues (20)
- Dockerfile not working HOT 7
- ModuleNotFoundError: No module named 'cyberbattle' HOT 2
- E: Unable to locate package python3.9 HOT 1
- Key for how the simulation is working HOT 1
- simplenv bidirectional? HOT 3
- Example agent doesn't work with conda install HOT 8
- python 3.11 not working - ValueError: mutable default <class 'cyberbattle.simulation.model.FirewallConfiguration'> HOT 9
- ValueError: mutable default <class 'cyberbattle.simulation.model.FirewallConfiguration'> for field firewall is not allowed: use default_factory HOT 3
- DQL still learning at evaluation time HOT 1
- Internal Node Id Dependent on Order of Action Execution HOT 1
- Unable to install dependencies HOT 12
- Action_space dimensions formally too large HOT 1
- Play the environment with other RL algorithms HOT 14
- A question about 'pretty_print_internal_action' HOT 2
- Making reward non-zero, what is the reason? HOT 1
- Something wrong with the node's last owned time and last reimaged time? HOT 1
- Consider designing defender as RL agent HOT 3
- init.sh doesn't work HOT 2
- How to setup machine/program for use HOT 5
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from cyberbattlesim.