cajcodes commited on
Commit
502a473
1 Parent(s): b1581fb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +28 -1
README.md CHANGED
@@ -39,6 +39,14 @@ The model was trained using a hybrid approach:
39
  - Target Update Frequency: Every 100 episodes
40
  - Number of Episodes: 50
41
 
 
 
 
 
 
 
 
 
42
  ## Usage
43
 
44
  To use this model, load the saved state dictionary and initialize the DQN with the same architecture. The model can then be used to navigate a floorplan and find the most efficient path to the target.
@@ -84,6 +92,21 @@ state = ... # Define your state here
84
  with torch.no_grad():
85
  action = model(torch.tensor(state, dtype=torch.float32).unsqueeze(0)).argmax().item()
86
  ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
87
  ## Evaluation
88
 
89
  The model was evaluated based on:
@@ -104,6 +127,10 @@ The model was evaluated based on:
104
 
105
  This project leverages the power of reinforcement learning combined with traditional pathfinding algorithms to navigate complex environments efficiently.
106
 
 
 
 
 
107
  ## Citation
108
 
109
  If you use this model in your research, please cite it as follows:
@@ -112,7 +139,7 @@ If you use this model in your research, please cite it as follows:
112
  author = {Christopher Jones},
113
  title = {Deep Q-Network for Floorplan Navigation},
114
  year = {2024},
115
- howpublished = {\url{https://huggingface.co/cajcodes/dqn-floorplan-finder}},
116
  note = {Accessed: YYYY-MM-DD}
117
  }
118
  ```
 
39
  - Target Update Frequency: Every 100 episodes
40
  - Number of Episodes: 50
41
 
42
+ ## Checkpoints
43
+
44
+ Checkpoints are saved during training for convenience:
45
+ - `checkpoint_11.pth.tar`: After 11 episodes
46
+ - `checkpoint_21.pth.tar`: After 21 episodes
47
+ - `checkpoint_31.pth.tar`: After 31 episodes
48
+ - `checkpoint_41.pth.tar`: After 41 episodes
49
+
50
  ## Usage
51
 
52
  To use this model, load the saved state dictionary and initialize the DQN with the same architecture. The model can then be used to navigate a floorplan and find the most efficient path to the target.
 
92
  with torch.no_grad():
93
  action = model(torch.tensor(state, dtype=torch.float32).unsqueeze(0)).argmax().item()
94
  ```
95
+ ## Training Script
96
+
97
+ The training script train.py is included in the repository for those who wish to reproduce the training process or continue training from a specific checkpoint.
98
+
99
+ ### Training Instructions
100
+ - Clone the repository.
101
+ - Ensure you have the necessary dependencies installed.
102
+ - Run the training script:
103
+ ```
104
+ bash
105
+ Copy code
106
+ python train.py
107
+ ```
108
+ To continue training from a checkpoint, modify the script to load the checkpoint before training.
109
+
110
  ## Evaluation
111
 
112
  The model was evaluated based on:
 
127
 
128
  This project leverages the power of reinforcement learning combined with traditional pathfinding algorithms to navigate complex environments efficiently.
129
 
130
+ ## License
131
+
132
+ This model is licensed under the Apache 2.0 License.
133
+
134
  ## Citation
135
 
136
  If you use this model in your research, please cite it as follows:
 
139
  author = {Christopher Jones},
140
  title = {Deep Q-Network for Floorplan Navigation},
141
  year = {2024},
142
+ howpublished = {\url{https://huggingface.co/cajcodes/dqn-floorplan-navigator}},
143
  note = {Accessed: YYYY-MM-DD}
144
  }
145
  ```