forked from louisnino/RLcode
-
Notifications
You must be signed in to change notification settings - Fork 0
/
tutorial_DQN.py
213 lines (173 loc) · 9.54 KB
/
tutorial_DQN.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
"""
Deep Q-Network Q(a, s)
-----------------------
TD Learning, Off-Policy, e-Greedy Exploration (GLIE).
Q(S, A) <- Q(S, A) + alpha * (R + lambda * Q(newS, newA) - Q(S, A))
delta_w = R + lambda * Q(newS, newA)
See David Silver RL Tutorial Lecture 5 - Q-Learning for more details.
Reference
----------
original paper: https://storage.googleapis.com/deepmind-media/dqn/DQNNaturePaper.pdf
EN: https://medium.com/emergent-future/simple-reinforcement-learning-with-tensorflow-part-0-q-learning-with-tables-and-neural-networks-d195264329d0#.5m3361vlw
CN: https://zhuanlan.zhihu.com/p/25710327
Note: Policy Network has been proved to be better than Q-Learning, see tutorial_atari_pong.py
Environment
-----------
# The FrozenLake v0 environment
https://gym.openai.com/envs/FrozenLake-v0
The agent controls the movement of a character in a grid world. Some tiles of
the grid are walkable, and others lead to the agent falling into the water.
Additionally, the movement direction of the agent is uncertain and only partially
depends on the chosen direction. The agent is rewarded for finding a walkable
path to a goal tile.
SFFF (S: starting point, safe)
FHFH (F: frozen surface, safe)
FFFH (H: hole, fall to your doom)
HFFG (G: goal, where the frisbee is located)
The episode ends when you reach the goal or fall in a hole. You receive a reward
of 1 if you reach the goal, and zero otherwise.
Prerequisites
--------------
tensorflow>=2.0.0a0
tensorlayer>=2.0.0
To run
-------
python tutorial_DQN.py --train/test
"""
import argparse
import time
import gym
import numpy as np
import tensorflow as tf
import tensorlayer as tl
# add arguments in command --train/test
# 关于argparase的应用,可以看看我这篇知乎专栏:
# 小段文讲清argparse模块基本用法[小番外]
# https://zhuanlan.zhihu.com/p/111010774
# 注意:原代码默认为test,我改为了train。
parser = argparse.ArgumentParser(description='Train or test neural net motor controller.')
parser.add_argument('--train', dest='train', action='store_true', default=True)
parser.add_argument('--test', dest='test', action='store_true', default=False)
args = parser.parse_args()
tl.logging.set_verbosity(tl.logging.DEBUG)
##################### hyper parameters ####################
lambd = .99 # 折扣率(decay factor)
e = 0.1 # epsilon-greedy算法参数,越大随机性越大,越倾向于探索行为。
num_episodes = 10000 # 迭代次数
render = False # 是否渲染游戏
running_reward = None
##################### DQN ##########################
## 把分类的数字表示,变成onehot表示。
# 例如有4类,那么第三类变为:[0,0,1,0]的表示。
def to_one_hot(i, n_classes=None):
a = np.zeros(n_classes, 'uint8') # 这里先按照分类数量构建一个全0向量
a[i] = 1 # 然后点亮需要onehot的位数。
return a
## Define Q-network q(a,s) that ouput the rewards of 4 actions by given state, i.e. Action-Value Function.
# encoding for state: 4x4 grid can be represented by one-hot vector with 16 integers.
def get_model(inputs_shape):
'''
定义Q网络模型:
1. 注意输入的shape和输出的shape
2. W_init和b_init是模型在初始化的时候,控制初始化参数的随机。该代码中用正态分布,均值0,方差0.01的方式初始化参数。
'''
ni = tl.layers.Input(inputs_shape, name='observation')
nn = tl.layers.Dense(4, act=None, W_init=tf.random_uniform_initializer(0, 0.01), b_init=None, name='q_a_s')(ni)
return tl.models.Model(inputs=ni, outputs=nn, name="Q-Network")
def save_ckpt(model): # save trained weights
'''
保存参数
'''
tl.files.save_npz(model.trainable_weights, name='dqn_model.npz')
def load_ckpt(model): # load trained weights
'''
加载参数
'''
tl.files.load_and_assign_npz(name='dqn_model.npz', network=model)
if __name__ == '__main__':
qnetwork = get_model([None, 16]) #定义inputshape[None,16]。16是state数量
qnetwork.train() #调用tensorlayer的时候,需要标注这个模型是否可以训练。(再次吐槽tenorlayers...)
train_weights = qnetwork.trainable_weights #模型的参数
optimizer = tf.optimizers.SGD(learning_rate=0.1) #定义优化器
env = gym.make('FrozenLake-v0') #定义环境
# ======开始训练=======
if args.train:
t0 = time.time()
for i in range(num_episodes):
## 重置环境初始状态
s = env.reset()
rAll = 0
for j in range(99): # 最多探索99步。因为环境状态比较少,99步一般也够探索到最终状态了。
if render: env.render()
## 把state放入network,计算Q值。
## 注意,这里先把state进行onehote处理,这里注意解释下什么是onehot
## 输出,这个状态下,所有动作的Q值,也就是说,是一个[None,4]大小的矩阵
allQ = qnetwork(np.asarray([to_one_hot(s, 16)], dtype=np.float32)).numpy()
# 在矩阵中找最大的Q值的动作
a = np.argmax(allQ, 1)
# e-Greedy:如果小于epsilon,就智能体随机探索。否则,就用最大Q值的动作。
if np.random.rand(1) < e:
a[0] = env.action_space.sample()
# 输入到环境,获得下一步的state,reward,done
s1, r, d, _ = env.step(a[0])
# 把new-state 放入,预测下一个state的**所有动作**的Q值。
Q1 = qnetwork(np.asarray([to_one_hot(s1, 16)], dtype=np.float32)).numpy()
##=======计算target=======
## 构建更新target:
# Q'(s,a) <- Q(s,a) + alpha(r + lambd * maxQ(s',a') - Q(s, a))
maxQ1 = np.max(Q1) # 下一个状态中最大Q值.
targetQ = allQ # 用allQ(现在状态的Q值)构建更新的target。因为只有被选择那个动作才会被更新到。
targetQ[0, a[0]] = r + lambd * maxQ1
## 利用自动求导 进行更新。
with tf.GradientTape() as tape:
_qvalues = qnetwork(np.asarray([to_one_hot(s, 16)], dtype=np.float32)) #把s放入到Q网络,计算_qvalues。
#_qvalues和targetQ的差距就是loss。这里衡量的尺子是mse
_loss = tl.cost.mean_squared_error(targetQ, _qvalues, is_mean=False)
# 同梯度带求导对网络参数求导
grad = tape.gradient(_loss, train_weights)
# 应用梯度到网络参数求导
optimizer.apply_gradients(zip(grad, train_weights))
# 累计reward,并且把s更新为newstate
rAll += r
s = s1
#更新epsilon,让epsilon随着迭代次数增加而减少。
#目的就是智能体越来越少进行“探索”
if d ==True:
e = 1. / ((i / 50) + 10)
break
## 这里的running_reward用于记载每一次更新的总和。为了能够更加看清变化,所以大部分是前面的。只有一部分是后面的。
running_reward = rAll if running_reward is None else running_reward * 0.99 + rAll * 0.01
# print("Episode [%d/%d] sum reward: %f running reward: %f took: %.5fs " % \
# (i, num_episodes, rAll, running_reward, time.time() - episode_time))
print('Episode: {}/{} | Episode Reward: {:.4f} | Running Average Reward: {:.4f} | Running Time: {:.4f}'\
.format(i, num_episodes, rAll, running_reward, time.time()-t0 ))
save_ckpt(qnetwork) # save model
##============这部分是正式游戏了========
# 这部分就不讲解了,和训练一样。只是少了epsilon-greedy。
if args.test:
t0 = time.time()
load_ckpt(qnetwork) # load model
for i in range(num_episodes):
## Reset environment and get first new observation
episode_time = time.time()
s = env.reset() # observation is state, integer 0 ~ 15
rAll = 0
for j in range(99): # step index, maximum step is 99
if render: env.render()
## Choose an action by greedily (with e chance of random action) from the Q-network
allQ = qnetwork(np.asarray([to_one_hot(s, 16)], dtype=np.float32)).numpy()
a = np.argmax(allQ, 1) # no epsilon, only greedy for testing
## Get new state and reward from environment
s1, r, d, _ = env.step(a[0])
rAll += r
s = s1
## Reduce chance of random action if an episode is done.
if d ==True:
#e = 1. / ((i / 50) + 10) # reduce e, GLIE: Greey in the limit with infinite Exploration
break
## Note that, the rewards here with random action
running_reward = rAll if running_reward is None else running_reward * 0.99 + rAll * 0.01
# print("Episode [%d/%d] sum reward: %f running reward: %f took: %.5fs " % \
# (i, num_episodes, rAll, running_reward, time.time() - episode_time))
print('Episode: {}/{} | Episode Reward: {:.4f} | Running Average Reward: {:.4f} | Running Time: {:.4f}'\
.format(i, num_episodes, rAll, running_reward, time.time()-t0 ))