You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
-`dueling::Bool` dueling structure for the q network default = true
124
124
-`recurrence::Bool = false` set to true to use DRQN, it will throw an error if you set it to false and pass a recurrent model.
125
+
-`evaluation_policy::Function = basic_evaluation` function use to evaluate the policy every `eval_freq` steps, the default is a rollout that return the undiscounted average reward
-`buffer_size::Int64` size of the experience replay buffer default = 1000
130
131
-`max_episode_length::Int64` maximum length of a training episode default = 100
131
132
-`train_start::Int64` number of steps used to fill in the replay buffer initially default = 200
132
-
-`save_freq::Int64` save the model every `save_freq` steps, default = 1000
133
-
-`evaluation_policy::Function = basic_evaluation` function use to evaluate the policy every `eval_freq` steps, the default is a rollout that return the undiscounted average reward
134
-
-`exploration_policy::Any = linear_epsilon_greedy(max_steps, eps_fraction, eps_end)` exploration strategy (default is epsilon greedy with linear decay)
135
133
-`rng::AbstractRNG` random number generator default = MersenneTwister(0)
136
134
-`logdir::String = ""` folder in which to save the model
135
+
-`save_freq::Int64` save the model every `save_freq` steps, default = 1000
136
+
-`log_freq::Int64` frequency at which to logg info default = 100
0 commit comments