Skip to content

Commit

Permalink
added the experiments
Browse files Browse the repository at this point in the history
  • Loading branch information
ferielamira1 committed Jan 17, 2024
1 parent 2688947 commit 5f09d2f
Show file tree
Hide file tree
Showing 68 changed files with 285 additions and 0 deletions.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
25 changes: 25 additions & 0 deletions abm/agent/experiments/exp_0/hyperparameters.txt
Original file line number Diff line number Diff line change
@@ -0,0 +1,25 @@
DQN Architecture and Hyperparameters:
DQNAgent:
- number of agents: 2
- state_size: self.v_field_res+ 1, action_size=3
- action_size: 3 [1: explore, 2: exploit, 3: relocate]
- replay_memory_capacity: 10,000
- batch_size: 128
- gamma: 0.99
- epsilon_start: 0.9
- tau: 0.005
- epsilon_decay: 1000
- epsilon_end: 0.05
- lr: 1e-4
- reward = collected / time

DQNetwork:
- input_size: [Specify the size of the input]
- output_size: [Specify the size of the output layer]

Training Process:
- Experience replay with a deque (max capacity: 10,000)
- Epsilon-greedy exploration
- Q-network trained with mini-batches (batch size: 128)
- Mean Squared Error (MSE) loss
- Target Q-network updated with soft update (tau: 0.005) every step
Binary file added abm/agent/experiments/exp_0/model_0.pth
Binary file not shown.
Binary file added abm/agent/experiments/exp_0/model_1.pth
Binary file not shown.
Binary file not shown.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
25 changes: 25 additions & 0 deletions abm/agent/experiments/exp_1/hyperparameters.txt
Original file line number Diff line number Diff line change
@@ -0,0 +1,25 @@
DQN Architecture and Hyperparameters:
DQNAgent:
- number of agents: 2
- state_size: self.v_field_res+ 1, action_size=3
- action_size: 3 [1: explore, 2: exploit, 3: relocate]
- replay_memory_capacity: 10,000
- batch_size: 128
- gamma: 0.99
- epsilon_start: 0.9
- tau: 0.005
- epsilon_decay: 1000
- epsilon_end: 0.05
- lr: 1e-5
- reward = collected / time

DQNetwork:
- input_size: [Specify the size of the input]
- output_size: [Specify the size of the output layer]

Training Process:
- Experience replay with a deque (max capacity: 10,000)
- Epsilon-greedy exploration
- Q-network trained with mini-batches (batch size: 128)
- Mean Squared Error (MSE) loss
- Target Q-network updated with soft update (tau: 0.005) every step
Binary file added abm/agent/experiments/exp_1/model_0.pth
Binary file not shown.
Binary file added abm/agent/experiments/exp_1/model_1.pth
Binary file not shown.
Binary file not shown.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
27 changes: 27 additions & 0 deletions abm/agent/experiments/exp_10/hyperparameters.txt
Original file line number Diff line number Diff line change
@@ -0,0 +1,27 @@
DQN-Architecture and Hyperparameters:

DQNAgent:
- number of agents: 3
- state_size: self.v_field_res+ 1, action_size=3
- action_size: 3 [1: explore, 2: exploit, 3: relocate]
- replay_memory_capacity: 10,000
- batch_size: 128
- gamma: 0.99
- epsilon_start: 0.9
- tau: 0.005
- epsilon_decay: 1000
- epsilon_end: 0.05
- lr: 1e-5 with scheduler
- if self.t!=0 reward= (0.2*ag.collected_r + 0.8*collective_reward) /self.t else reward=0
where collective_reward = sum of ag.collected_r / (self.t*len(agents))

DQNetwork:
- input_size: [Specify the size of the input]
- output_size: [Specify the size of the output layer]

Training Process:
- Experience replay with a deque (max capacity: 10,000)
- Epsilon-greedy exploration
- Q-network trained with mini-batches (batch size: 128)
- Mean Squared Error (MSE) loss
- Target Q-network updated with soft update (tau: 0.005) every 50 iterations
Binary file added abm/agent/experiments/exp_10/model_0.pth
Binary file not shown.
Binary file added abm/agent/experiments/exp_10/model_1.pth
Binary file not shown.
Binary file added abm/agent/experiments/exp_10/model_2.pth
Binary file not shown.
Binary file not shown.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
26 changes: 26 additions & 0 deletions abm/agent/experiments/exp_2/hyperparameters.txt
Original file line number Diff line number Diff line change
@@ -0,0 +1,26 @@
DQN Architecture and Hyperparameters:
DQNAgent:
- number of agents: 1
- state_size: self.v_field_res+ 1, action_size=3
- action_size: 3 [1: explore, 2: exploit, 3: relocate]
- replay_memory_capacity: 10,000
- batch_size: 128
- gamma: 0.99
- epsilon_start: 0.9
- tau: 0.005
- epsilon_decay: 1000
- epsilon_end: 0.05
- lr: 1e-5 with scheduler
- reward = collected / time


DQNetwork:
- input_size: [Specify the size of the input]
- output_size: [Specify the size of the output layer]

Training Process:
- Experience replay with a deque (max capacity: 10,000)
- Epsilon-greedy exploration
- Q-network trained with mini-batches (batch size: 128)
- Mean Squared Error (MSE) loss
- Target Q-network updated with soft update (tau: 0.005) every step
Binary file added abm/agent/experiments/exp_2/model_0.pth
Binary file not shown.
Binary file not shown.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
26 changes: 26 additions & 0 deletions abm/agent/experiments/exp_3/hyperparameters.txt
Original file line number Diff line number Diff line change
@@ -0,0 +1,26 @@
DQN Architecture and Hyperparameters:
DQNAgent:
- number of agents: 1
- state_size: self.v_field_res+ 1, action_size=3
- action_size: 3 [1: explore, 2: exploit, 3: relocate]
- replay_memory_capacity: 10,000
- batch_size: 128
- gamma: 0.99
- epsilon_start: 0.9
- tau: 0.005
- epsilon_decay: 1000
- epsilon_end: 0.05
- lr: 1e-5 with scheduler
- reward = collected / time


DQNetwork:
- input_size: [Specify the size of the input]
- output_size: [Specify the size of the output layer]

Training Process:
- Experience replay with a deque (max capacity: 10,000)
- Epsilon-greedy exploration
- Q-network trained with mini-batches (batch size: 128)
- Mean Squared Error (MSE) loss
- Target Q-network updated with soft update (tau: 0.005) every 50 iterations
Binary file added abm/agent/experiments/exp_3/model_0.pth
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
25 changes: 25 additions & 0 deletions abm/agent/experiments/exp_4/hyperparameters.txt
Original file line number Diff line number Diff line change
@@ -0,0 +1,25 @@
DQN Architecture and Hyperparameters:
DQNAgent:
- number of agents: 1
- state_size: self.v_field_res+ 1, action_size=3
- action_size: 3 [1: explore, 2: exploit, 3: relocate]
- replay_memory_capacity: 10,000
- batch_size: 128
- gamma: 0.99
- epsilon_start: 0.9
- tau: 0.005
- epsilon_decay: 1000
- epsilon_end: 0.05
- lr: 1e-5 with scheduler
- reward = 1 if resource is exploited, 0 otherwise

DQNetwork:
- input_size: [Specify the size of the input]
- output_size: [Specify the size of the output layer]

Training Process:
- Experience replay with a deque (max capacity: 10,000)
- Epsilon-greedy exploration
- Q-network trained with mini-batches (batch size: 128)
- Mean Squared Error (MSE) loss
- Target Q-network updated with soft update (tau: 0.005) every 50 iterations
Binary file added abm/agent/experiments/exp_4/img.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Binary file added abm/agent/experiments/exp_4/model_0.pth
Binary file not shown.
Binary file added abm/agent/experiments/exp_4/model_1.pth
Binary file not shown.
Binary file not shown.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
26 changes: 26 additions & 0 deletions abm/agent/experiments/exp_5/hyperparameters.txt
Original file line number Diff line number Diff line change
@@ -0,0 +1,26 @@
DQN Architecture and Hyperparameters:
DQNAgent:
- number of agents: 2
- state_size: self.v_field_res+ 1, action_size=3
- action_size: 3 [1: explore, 2: exploit, 3: relocate]
- replay_memory_capacity: 10,000
- batch_size: 128
- gamma: 0.99
- epsilon_start: 0.9
- tau: 0.005
- epsilon_decay: 1000
- epsilon_end: 0.05
- lr: 1e-5 with scheduler
- o if not exploit else reward = collected / time


DQNetwork:
- input_size: [Specify the size of the input]
- output_size: [Specify the size of the output layer]

Training Process:
- Experience replay with a deque (max capacity: 10,000)
- Epsilon-greedy exploration
- Q-network trained with mini-batches (batch size: 128)
- Mean Squared Error (MSE) loss
- Target Q-network updated with soft update (tau: 0.005) every 50 iterations
Binary file added abm/agent/experiments/exp_5/model_0.pth
Binary file not shown.
Binary file added abm/agent/experiments/exp_5/model_1.pth
Binary file not shown.
Binary file not shown.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
26 changes: 26 additions & 0 deletions abm/agent/experiments/exp_6/hyperparameters.txt
Original file line number Diff line number Diff line change
@@ -0,0 +1,26 @@
DQN-Architecture and Hyperparameters:

DQNAgent:
- number of agents: 2
- state_size: self.v_field_res+ 1, action_size=3
- action_size: 3 [1: explore, 2: exploit, 3: relocate]
- replay_memory_capacity: 10,000
- batch_size: 128
- gamma: 0.99
- epsilon_start: 0.9
- tau: 0.005
- epsilon_decay: 1000
- epsilon_end: 0.05
- lr: 1e-5 with scheduler
- if self.t!=0 reward= ag.collected_r /self.t else reward=0

DQNetwork:
- input_size: [Specify the size of the input]
- output_size: [Specify the size of the output layer]

Training Process:
- Experience replay with a deque (max capacity: 10,000)
- Epsilon-greedy exploration
- Q-network trained with mini-batches (batch size: 128)
- Mean Squared Error (MSE) loss
- Target Q-network updated with soft update (tau: 0.005) every 50 iterations
Binary file added abm/agent/experiments/exp_6/model_0.pth
Binary file not shown.
Binary file added abm/agent/experiments/exp_6/model_1.pth
Binary file not shown.
Binary file not shown.
Binary file not shown.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
26 changes: 26 additions & 0 deletions abm/agent/experiments/exp_7/hyperparameters.txt
Original file line number Diff line number Diff line change
@@ -0,0 +1,26 @@
DQN-Architecture and Hyperparameters:

DQNAgent:
- number of agents: 3
- state_size: self.v_field_res+ 1, action_size=3
- action_size: 3 [1: explore, 2: exploit, 3: relocate]
- replay_memory_capacity: 10,000
- batch_size: 128
- gamma: 0.99
- epsilon_start: 0.9
- tau: 0.005
- epsilon_decay: 1000
- epsilon_end: 0.05
- lr: 1e-5 with scheduler
- if self.t!=0 reward= ag.collected_r /self.t else reward=0

DQNetwork:
- input_size: [Specify the size of the input]
- output_size: [Specify the size of the output layer]

Training Process:
- Experience replay with a deque (max capacity: 10,000)
- Epsilon-greedy exploration
- Q-network trained with mini-batches (batch size: 128)
- Mean Squared Error (MSE) loss
- Target Q-network updated with soft update (tau: 0.005) every 50 iterations
Binary file added abm/agent/experiments/exp_7/model_0.pth
Binary file not shown.
Binary file added abm/agent/experiments/exp_7/model_1.pth
Binary file not shown.
Binary file added abm/agent/experiments/exp_7/model_2.pth
Binary file not shown.
Binary file not shown.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
26 changes: 26 additions & 0 deletions abm/agent/experiments/exp_8/hyperparameters.txt
Original file line number Diff line number Diff line change
@@ -0,0 +1,26 @@
DQN-Architecture and Hyperparameters:

DQNAgent:
- number of agents: 3
- state_size: self.v_field_res+ 1, action_size=3
- action_size: 3 [1: explore, 2: exploit, 3: relocate]
- replay_memory_capacity: 10,000
- batch_size: 128
- gamma: 0.99
- epsilon_start: 0.9
- tau: 0.005
- epsilon_decay: 1000
- epsilon_end: 0.05
- lr: 1e-5 with scheduler
- if self.t!=0 reward= ag.collected_r + ag.collective_reward /self.t else reward=0

DQNetwork:
- input_size: [Specify the size of the input]
- output_size: [Specify the size of the output layer]

Training Process:
- Experience replay with a deque (max capacity: 10,000)
- Epsilon-greedy exploration
- Q-network trained with mini-batches (batch size: 128)
- Mean Squared Error (MSE) loss
- Target Q-network updated with soft update (tau: 0.005) every 50 iterations
Binary file added abm/agent/experiments/exp_8/model_0.pth
Binary file not shown.
Binary file added abm/agent/experiments/exp_8/model_1.pth
Binary file not shown.
Binary file added abm/agent/experiments/exp_8/model_2.pth
Binary file not shown.
Binary file not shown.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
27 changes: 27 additions & 0 deletions abm/agent/experiments/exp_9/hyperparameters.txt
Original file line number Diff line number Diff line change
@@ -0,0 +1,27 @@
DQN-Architecture and Hyperparameters:

DQNAgent:
- number of agents: 3
- state_size: self.v_field_res+ 1, action_size=3
- action_size: 3 [1: explore, 2: exploit, 3: relocate]
- replay_memory_capacity: 10,000
- batch_size: 128
- gamma: 0.99
- epsilon_start: 0.9
- tau: 0.005
- epsilon_decay: 1000
- epsilon_end: 0.05
- lr: 1e-5 with scheduler
- if self.t!=0 reward= (0.2*ag.collected_r + 0.8*collective_reward) /self.t else reward=0
where collective_reward = sum of ag.collected_r / (self.t*len(agents))

DQNetwork:
- input_size: [Specify the size of the input]
- output_size: [Specify the size of the output layer]

Training Process:
- Experience replay with a deque (max capacity: 10,000)
- Epsilon-greedy exploration
- Q-network trained with mini-batches (batch size: 128)
- Mean Squared Error (MSE) loss
- Target Q-network updated with soft update (tau: 0.005) every 50 iterations
Binary file added abm/agent/experiments/exp_9/model_0.pth
Binary file not shown.
Binary file added abm/agent/experiments/exp_9/model_1.pth
Binary file not shown.
Binary file added abm/agent/experiments/exp_9/model_2.pth
Binary file not shown.
Binary file not shown.

0 comments on commit 5f09d2f

Please sign in to comment.