gomoku / DI-engine /dizoo /smac /config /smac_MMM_collaq_config.py
zjowowen's picture
init space
079c32c
raw
history blame
2.67 kB
from easydict import EasyDict
agent_num = 10
collector_env_num = 16
evaluator_env_num = 8
main_config = dict(
exp_name='smac_MMM_collaq_seed0',
env=dict(
map_name='MMM',
difficulty=7,
reward_only_positive=True,
mirror_opponent=False,
agent_num=agent_num,
collector_env_num=collector_env_num,
evaluator_env_num=evaluator_env_num,
shared_memory=False,
stop_value=0.999,
n_evaluator_episode=32,
obs_alone=True,
manager=dict(
shared_memory=False,
reset_timeout=6000,
),
),
policy=dict(
model=dict(
agent_num=agent_num,
obs_shape=186,
alone_obs_shape=114,
global_obs_shape=290,
action_shape=16,
hidden_size_list=[128],
attention=False,
# obs_shape = move_feature(4) + enemy_feats(enemy_feat_dim*enemy_num)
# + ally_feats(ally_feat_dim*ally_num) + own_feats + agent_id_feats (agent_num)
# please see the function of get_obs_agent in smac_env.py
self_feature_range=[156, 176],
ally_feature_range=[84, 156],
attention_size=32,
mixer=True,
lstm_type='gru',
dueling=False,
),
learn=dict(
update_per_collect=20,
batch_size=32,
learning_rate=0.0005,
clip_value=5,
double_q=True,
target_update_theta=0.008,
discount_factor=0.95,
collaq_loss_weight=1.0,
),
collect=dict(
n_episode=32,
unroll_len=10,
env_num=collector_env_num,
),
eval=dict(env_num=evaluator_env_num, evaluator=dict(eval_freq=100, )),
other=dict(
eps=dict(
type='linear',
start=1,
end=0.05,
decay=10000,
),
replay_buffer=dict(
replay_buffer_size=15000,
# (int) The maximum reuse times of each data
max_reuse=1e+9,
max_staleness=1e+9,
),
),
),
)
main_config = EasyDict(main_config)
create_config = dict(
env=dict(
type='smac',
import_names=['dizoo.smac.envs.smac_env'],
),
env_manager=dict(type='subprocess'),
policy=dict(type='collaq'),
collector=dict(type='episode', get_train_sample=True),
)
create_config = EasyDict(create_config)
if __name__ == '__main__':
from ding.entry import serial_pipeline
serial_pipeline((main_config, create_config), seed=0)