fix (xtuner): change import form and simplify neural network

This commit is contained in:
wangtq
2021-06-07 14:56:07 +08:00
parent 3d8e90e705
commit 120feef703

View File

@ -13,10 +13,10 @@ MERCHANTABILITY OR FIT FOR A PARTICULAR PURPOSE.
See the Mulan PSL v2 for more details.
"""
from keras import Model
from keras import Sequential
from keras.layers import Flatten, Dense, Activation, Input, Concatenate
from keras.optimizers import Adam
from tensorflow.keras import Model
from tensorflow.keras import Sequential
from tensorflow.keras.layers import Flatten, Dense, Activation, Input, Concatenate
from tensorflow.keras.optimizers import Adam
from rl.agents import DDPGAgent
from rl.agents import DQNAgent
from rl.memory import SequentialMemory
@ -47,8 +47,6 @@ class RLAgent:
model.add(Activation('relu'))
model.add(Dense(16))
model.add(Activation('relu'))
model.add(Dense(16))
model.add(Activation('relu'))
model.add(Dense(nb_actions, activation='linear'))
# build alg
@ -70,8 +68,6 @@ class RLAgent:
actor.add(Activation('relu'))
actor.add(Dense(16))
actor.add(Activation('relu'))
actor.add(Dense(16))
actor.add(Activation('relu'))
actor.add(Dense(nb_actions))
actor.add(Activation('sigmoid'))
@ -84,8 +80,6 @@ class RLAgent:
x = Activation('relu')(x)
x = Dense(32)(x)
x = Activation('relu')(x)
x = Dense(32)(x)
x = Activation('relu')(x)
x = Dense(1)(x)
x = Activation('linear')(x)
critic = Model(inputs=[action_input, observation_input], outputs=x)