-
Notifications
You must be signed in to change notification settings - Fork 1
/
Copy pathconfig_params.py
65 lines (55 loc) · 1.42 KB
/
config_params.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
import os
gridworld_params = {
'size': 15,
'max_episode_steps': 60,
}
manipulation_params = {
'task': 'chest_pick_and_place',
'gripper': 'parallel_jaw',
'num_block': 1,
'distance_threshold': 0.03,
'max_episode_steps': 50,
'task_decomposition': False,
'binary_reward': True,
}
agent_params = {
'hindsight': True,
'her_sampling_strategy': 'future',
'prioritised': False,
'memory_capacity': int(1e6),
'actor_learning_rate': 0.001,
'critic_learning_rate': 0.001,
'update_interval': 1,
'batch_size': 128,
'optimization_steps': 40,
'tau': 0.05,
'discount_factor': 0.98,
'clip_value': 50,
'observation_type': 'observation',
'discard_time_limit': False,
'terminate_on_achieve': True,
'observation_normalization': True,
# TA2 params
'abstract_demonstration': False,
'abstract_demonstration_eta': 0.75,
'adaptive_exploration': False,
'adaptive_exploration_tau': 0.3,
# DQN params
'e_greed_decay': 5e5,
# DDPG params
'Q_weight_decay': 0.0,
'exploration_epsilon': 0.2,
'exploration_sigma': 0.05,
# SAC params
'alpha': 1.0,
'actor_update_interval': 2,
'critic_target_update_interval': 2,
'curriculum': False,
'training_epochs': 51,
'training_cycles': 50,
'training_episodes': 16,
'testing_gap': 1,
'testing_episodes': 30,
'saving_gap': 50,
'cuda_device_id': 0
}