赞
踩
CartPole问题:黑色小车上面支撑的一个连接杆,连杆会自由摆动,我们需要控制黑色小车,通过控制小车左右移动,保持连杆的平衡。
该问题的动作空间是离散的且有限的,只有两种执行动作(0或1),但是该问题的状态空间是一个连续空间,且每个状态是一个四维向量。
执行动作:
### Action Space
The action is a `ndarray` with shape `(1,)` which can take values `{0, 1}` indicating the direction of the fixed force the cart is pushed with.
| Num | Action |
|-----|------------------------|
| 0 | Push cart to the left |
| 1 | Push cart to the right |
状态空间:
### Observation Space
The observation is a `ndarray` with shape `(4,)` with the values corresponding to the following positions and velocities:
| Num | Observation | Min | Max |
|-----|-----------------------|----------------------|--------------------|
| 0 | Cart Position | -4.8 | 4.8 |
| 1 | Cart Velocity | -Inf | Inf |
| 2 | Pole Angle | ~ -0.418 rad (-24°) | ~ 0.418 rad (24°) |
| 3 | Pole Angular Velocity | -Inf | Inf |
**Note:** While the ranges above denote the possible values for observation space of each element, it is not reflective of the allowed values of the state space in an unterminated episode. Particularly:
- The cart x-position (index 0) can be take values between `(-4.8, 4.8)`, but the episode terminates if the cart leaves the `(-2.4, 2.4)` range.
- The pole angle can be observed between `(-.418, .418)` radians (or **±24°**), but the episode terminates if the pole angle is not in the range `(-.2095, .2095)` (or **±12°**)
论文地址: Playing Atari with Deep Reinforcement Learning(https://arxiv.org/pdf/1312.5602.pdf)
版本信息
# -*- coding: utf-8 -*- import random import gym # 版本0.23.1 import numpy as np from collections import deque from tensorflow.keras.models import Sequential from tensorflow.keras.layers import Dense from tensorflow.keras.optimizers import Adam EPISODES = 1000 class DQNAgent: def __init__(self, state_size, action_size): self.state_size = state_size self.action_size = action_size self.memory = deque(maxlen=2000) # 记忆体使用队列实现,队列满后根据插入顺序自动删除老数据 self.gamma = 0.95 # discount rate self.epsilon = 0.4 # exploration rate self.epsilon_min = 0.01 self.epsilon_decay = 0.995 self.learning_rate = 0.001 self.model = self._build_model() # 可视化MLP结构 # plot_model(self.model, to_file='dqn-cartpole-v0-mlp.png', show_shapes=False) def _build_model(self): # Neural Net for Deep-Q learning Model model = Sequential() # 顺序模型,搭建神经网络(多层感知机) model.add(Dense(24, input_dim=self.state_size, activation='relu')) model.add(Dense(24, activation='relu')) model.add(Dense(self.action_size, activation='linear')) model.compile(loss='mse',optimizer=Adam(lr=self.learning_rate)) # 指定损失函数以及优化器 return model # 在记忆体(经验回放池)中保存具体某一时刻的当前状态信息 def remember(self, state, action, reward, next_state, done): # 当前状态、动作、奖励、下一个状态、是否结束 self.memory.append((state, action, reward, next_state, done)) # 根据模型预测结果返回动作 def act(self, state): if np.random.rand() <= self.epsilon: # 如果随机数(0-1之间)小于epsilon,则随机返回一个动作 return random.randrange(self.action_size) # 随机返回动作0或1 act_values = self.model.predict(state) # eg:[[0.35821578 0.11153378]] # print("model.predict act_values:",act_values) return np.argmax(act_values[0]) # returns action 返回价值最大的 # 记忆回放,训练神经网络模型 def replay(self, batch_size): minibatch = random.sample(self.memory, batch_size) for state, action, reward, next_state, done in minibatch: target = reward if not done: # 没有结束 target = (reward + self.gamma * np.amax(self.model.predict(next_state)[0])) target_f = self.model.predict(state) target_f[0][action] = target self.model.fit(state, target_f, epochs=1, verbose=0) # 训练神经网络 # 加载模型权重文件 def load(self, name): self.model.load_weights(name) # 保存模型 (参数:filepath) def save(self, name): self.model.save_weights(name) if __name__ == "__main__": env = gym.make('CartPole-v0') print(env.action_space) print(env.observation_space) state_size = env.observation_space.shape[0] action_size = env.action_space.n print("state_size:",state_size) # 4 print("action_size:",action_size) # 2 agent = DQNAgent(state_size, action_size) done = False batch_size = 32 avg=0 for e in range(EPISODES): # 循环学习次数,每次学习都需要初始化环境 state = env.reset() # 环境初始化,返回state例如[-0.1240581 -1.3752123 0.18474717 2.2276523 ] state = np.reshape(state, [1, state_size]) # 扩展维度(用于神经网络训练) # [[-0.1240581 -1.3752123 0.18474717 2.2276523 ]] for time in range(500): # 每次学习的步长为500 env.render() # 渲染可视化图像 # print(state) action = agent.act(state) # 根据模型预测结果返回动作 # print(action) # 0或者1 next_state, reward, done, _ = env.step(action) # 返回下一个状态、奖励、以及是否结束游戏(当摆杆出界或倾斜浮动等状态信息不符要求或步长大于内置值时结束游戏) reward = reward if not done else -10 # 结束游戏时,设置奖励为-10 next_state = np.reshape(next_state, [1, state_size]) agent.remember(state, action, reward, next_state, done) # 放入记忆体 state = next_state if done: print("episode: {}/{}, score(time): {}" .format(e, EPISODES, time)) avg += time break # 定期检查记忆大小,进行记忆回放 if len(agent.memory) > batch_size: agent.replay(batch_size) print("Avg score:{}".format(avg/1000))
前期,智能体(Agent)控制小车移动只能玩10秒左右
episode: 0/1000, score(time): 8 episode: 1/1000, score(time): 10 episode: 2/1000, score(time): 8 episode: 3/1000, score(time): 14 episode: 4/1000, score(time): 8 episode: 5/1000, score(time): 10 episode: 6/1000, score(time): 9 episode: 7/1000, score(time): 11 episode: 8/1000, score(time): 12 episode: 9/1000, score(time): 9 episode: 10/1000, score(time): 15 episode: 11/1000, score(time): 9 episode: 12/1000, score(time): 13 episode: 13/1000, score(time): 11 episode: 14/1000, score(time): 9 episode: 15/1000, score(time): 12 episode: 16/1000, score(time): 9 episode: 17/1000, score(time): 11
通过神经网络模型的不断训练…
episode: 267/1000, score(time): 155
episode: 268/1000, score(time): 188
episode: 269/1000, score(time): 100
episode: 270/1000, score(time): 136
episode: 271/1000, score(time): 126
episode: 272/1000, score(time): 155
episode: 273/1000, score(time): 179
episode: 274/1000, score(time): 104
episode: 275/1000, score(time): 111
episode: 276/1000, score(time): 199
episode: 277/1000, score(time): 128
episode: 278/1000, score(time): 199
可以看到智能体(Agent)的游戏水平不断提高
Copyright © 2003-2013 www.wpsshop.cn 版权所有,并保留所有权利。