-
Notifications
You must be signed in to change notification settings - Fork 9
/
prioritized_memory.py
54 lines (43 loc) · 1.56 KB
/
prioritized_memory.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
import random
import numpy as np
from SumTree import SumTree
class Memory: # stored as ( s, a, r, s_ ) in SumTree
e = 0.01
a = 0.6
beta = 0.4
beta_increment_per_sampling = 0.001
def __init__(self, capacity):
self.tree = SumTree(capacity)
self.capacity = capacity
def _get_priority(self, error):
return (np.abs(error) + self.e) ** self.a
def add(self, error, state, action, reward, next_state):
p = self._get_priority(error)
self.tree.add(p, state, action, reward, next_state)
def sample(self, n):
states = []
actions = []
rewards = []
next_states = []
idxs = []
segment = self.tree.total() / n
priorities = []
self.beta = np.min([1., self.beta + self.beta_increment_per_sampling])
for i in range(n):
a = segment * i
b = segment * (i + 1)
s = random.uniform(a, b)
(idx, p, state, action, reward, next_state) = self.tree.get(s)
priorities.append(p)
states.append(state)
actions.append(action)
rewards.append(reward)
next_states.append(next_state)
idxs.append(idx)
sampling_probabilities = priorities / self.tree.total()
is_weight = np.power(self.tree.n_entries * sampling_probabilities, -self.beta)
is_weight /= is_weight.max()
return states, actions, rewards, next_states, idxs, is_weight
def update(self, idx, error):
p = self._get_priority(error)
self.tree.update(idx, p)