-
Notifications
You must be signed in to change notification settings - Fork 2
Expand file tree
/
Copy pathrandomag.py
More file actions
executable file
·52 lines (38 loc) · 870 Bytes
/
randomag.py
File metadata and controls
executable file
·52 lines (38 loc) · 870 Bytes
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
import numpy as np
from common.utils import default_config, make_env
render = False
normalize_inputs = True
config = default_config()
env = make_env(config, normalize_inputs)
n_agent = env.n_agent
nD = env.nD
n_episode = env.n_episode
max_steps = env.max_steps
n_actions = env.n_actions
i_episode = 0
while i_episode<n_episode:
i_episode+=1
score=0
steps=0
su=[0.]*nD
su = np.array(su)
obs = env.reset()
done = False
while steps<max_steps and not done:
steps+=1
action=[]
for i in range(n_agent):
action.append(np.random.choice(range(n_actions)))
obs, rewards, done = env.step(action)
su+=np.array(rewards)
score += sum(rewards)
# if steps % 100 == 0:
# print(steps)
if render:
env.render()
print(i_episode)
print(score/max_steps)
print(su)
uti = np.array(su)/max_steps
print(env.rinfo.flatten())
env.end_episode()