diff --git a/README.md b/README.md
index 3f47a00..741b2d3 100644
--- a/README.md
+++ b/README.md
@@ -35,7 +35,6 @@ Firstly, each algorithm is implemented strictly according to the original paper
- [Overview of Algorithms](#overview-of-algorithms)
- [Supported Environments: Safety-Gymnasium](#supported-environments-safety-gymnasium)
- - [Gymnasium-based Environments](#gymnasium-based-environments)
- [Isaac Gym-based Environments](#isaac-gym-based-environments)
- [Selected Tasks](#selected-tasks)
- [Pre-requisites](#pre-requisites)
@@ -75,7 +74,6 @@ Here we provide a table of Safe RL algorithms that the benchmark includes.
For more details, please refer to [Safety-Gymnasium](https://github.com/PKU-Alignment/safety-gymnasium).
-### Gymnasium-based Environments
@@ -145,7 +143,12 @@ For more details, please refer to [Safety-Gymnasium](https://github.com/PKU-Alig
-**note**: Safe velocity tasks support both single-agent and multi-agent algorithms, while safe navigation tasks only support single-agent algorithms currently.
+**note**:
+
+- **Safe Velocity** and **Safe Isaac Gym** tasks support both single-agent and multi-agent algorithms.
+- **Safe Navigation** tasks support single-agent algorithms.
+- **Safe MultiGoal** tasks support multi-agent algorithms.
+- **Safe Isaac Gym** tasks do not support evaluation after training yet.
### Isaac Gym-based Environments
diff --git a/safepo/evaluate.py b/safepo/evaluate.py
index dee85cb..2818ea7 100644
--- a/safepo/evaluate.py
+++ b/safepo/evaluate.py
@@ -137,7 +137,7 @@ def single_runs_eval(eval_dir, eval_episodes):
config_path = eval_dir + '/config.json'
config = json.load(open(config_path, 'r'))
env = config['task'] if 'task' in config.keys() else config['env_name']
- if env in multi_agent_velocity_map.keys():
+ if env in multi_agent_velocity_map.keys() or env in multi_agent_goal_tasks:
reward, cost = eval_multi_agent(eval_dir, eval_episodes)
else:
reward, cost = eval_single_agent(eval_dir, eval_episodes)