Skip to content

Commit 798dd9e

Browse files
Update README.md
1 parent 0200708 commit 798dd9e

File tree

1 file changed

+16
-16
lines changed

1 file changed

+16
-16
lines changed

Diff for: README.md

+16-16
Original file line numberDiff line numberDiff line change
@@ -63,65 +63,65 @@ Codes (it'll send you to appropriate folder):
6363
<br/>
6464
Experimental results (figures and tables on this page):
6565

66-
* <a href="#RL Q-Learning Environment-1. Experimental results">RL Q-Learning Environment-1. Experimental results</a>
67-
* <a href="#Q-learning algorithm resulted chart for the environment-1">Q-learning algorithm resulted chart for the environment-1</a>
68-
* <a href="#Final Q-table with values from the final shortest route for environment-1">Final Q-table with values from the final shortest route for environment-1</a>
69-
* <a href="#RL Q-Learning Environment-2. Experimental results">RL Q-Learning Environment-2. Experimental results</a>
70-
* <a href="#Q-learning algorithm resulted chart for the environment-2">Q-learning algorithm resulted chart for the environment-2</a>
71-
* <a href="#Final Q-table with values from the final shortest route for environment-1">Final Q-table with values from the final shortest route for environment-1</a>
72-
* <a href="#RL Q-Learning Environment-3. Experimental results">RL Q-Learning Environment-3. Experimental results</a>
73-
* <a href="#Comparison analysis of Q-Learning and Sarsa algorithms">Comparison analysis of Q-Learning and Sarsa algorithms</a>
66+
* [RL Q-Learning Environment-1. Experimental results](#rl-q-learning-environment-1-experimental-results)
67+
* [Q-learning algorithm resulted chart for the environment-1](#q-learning-algorithm-resulted-chart-for-the-environment-1)
68+
* [Final Q-table with values from the final shortest route for environment-1](#final-q-table-with-values-from-the-final-shortest-route-for-environment-1)
69+
* [RL Q-Learning Environment-2. Experimental results](#rl-q-learning-environment-2-experimental-results)
70+
* [Q-learning algorithm resulted chart for the environment-2](#q-learning-algorithm-resulted-chart-for-the-environment-2)
71+
* [Final Q-table with values from the final shortest route for environment-1](#final-q-table-with-values-from-the-final-shortest-route-for-environment-1)
72+
* [RL Q-Learning Environment-3. Experimental results](#rl-q-learning-environment-3-experimental-results)
73+
* [Comparison analysis of Q-Learning and Sarsa algorithms](#comparison-analysis-of-q-learning-and-sarsa-algorithms)
7474

7575
<br/>
7676

77-
### <a name="RL Q-Learning Environment-1. Experimental results">RL Q-Learning Environment-1. Experimental results</a>
77+
### <a id="rl-q-learning-environment-1-experimental-results">RL Q-Learning Environment-1. Experimental results</a>
7878
Environment-1 with mobile robot, goal and obstacles
7979

8080
<img src="images/Environment-1.gif" alt="RL_Q-Learning_E-1" width=362 height=391> <img src="images/Environment-1.png" alt="RL_Q-Learning_E-1" width=362 height=391>
8181

8282
<br/>
8383

84-
### <a name="Q-learning algorithm resulted chart for the environment-1">Q-learning algorithm resulted chart for the environment-1</a>
84+
### <a id="q-learning-algorithm-resulted-chart-for-the-environment-1">Q-learning algorithm resulted chart for the environment-1</a>
8585
Represents number of episodes via number of steps and number of episodes via cost for each episode
8686

8787
![RL_Q-Learning_C-1](images/Charts-1.png)
8888

8989
<br/>
9090

91-
### <a name="Final Q-table with values from the final shortest route for environment-1">Final Q-table with values from the final shortest route for environment-1</a>
91+
### <a id="final-q-table-with-values-from-the-final-shortest-route-for-environment-1">Final Q-table with values from the final shortest route for environment-1</a>
9292
![RL_Q-Learning_T-1](images/Q-Table-E-1.png)
9393
<br/>Looking at the values of the table we can see the decision for the next action made by agent (mobile robot). The sequence of final actions to reach the goal after the Q-table is filled with knowledge is the following: *down-right-down-down-down-right-down-right-down-right-down-down-right-right-up-up.*
9494
<br/>During the experiment with Q-learning algorithm the found shortest route to reach the goal for the environment-1 consist of 16 steps and the found longest rout to reach the goal consists of 185 steps.
9595

9696
<br/>
9797

98-
### <a name="RL Q-Learning Environment-2. Experimental results">RL Q-Learning Environment-2. Experimental results</a>
98+
### <a id="rl-q-learning-environment-2-experimental-results">RL Q-Learning Environment-2. Experimental results</a>
9999
Bigger environment-2 with more obstacles
100100

101101
![RL_Q-Learning_E-2](images/Environment-2.png)
102102

103103
<br/>
104104

105-
### <a name="Q-learning algorithm resulted chart for the environment-2">Q-learning algorithm resulted chart for the environment-2</a>
105+
### <a id="q-learning-algorithm-resulted-chart-for-the-environment-2">Q-learning algorithm resulted chart for the environment-2</a>
106106
Represents number of episodes via number of steps and number of episodes via cost for each episode
107107

108108
![RL_Q-Learning_C-2](images/Charts-2.png)
109109

110110
<br/>
111111

112-
### <a name="Final Q-table with values from the final shortest route for environment-1">Final Q-table with values from the final shortest route for environment-1</a>
112+
### <a id="final-q-table-with-values-from-the-final-shortest-route-for-environment-1">Final Q-table with values from the final shortest route for environment-1</a>
113113
![RL_Q-Learning_T-2](images/Q-Table-E-2.png)
114114

115115
<br/>
116116

117-
### <a name="RL Q-Learning Environment-3. Experimental results">RL Q-Learning Environment-3. Experimental results</a>
117+
### <a id="rl-q-learning-environment-3-experimental-results">RL Q-Learning Environment-3. Experimental results</a>
118118
Super complex environment-3 with a lot of obstacles
119119

120120
![RL_Q-Learning_E-3](images/Environment-3.png)
121121

122122
<br/>
123123

124-
### <a name="Comparison analysis of Q-Learning and Sarsa algorithms">Comparison analysis of Q-Learning and Sarsa algorithms</a>
124+
### <a id="comparison-analysis-of-q-learning-and-sarsa-algorithms">Comparison analysis of Q-Learning and Sarsa algorithms</a>
125125
![RQ-learning_via_Sarsa](images/Q-learning_via_Sarsa.png)
126126

127127
<br/>

0 commit comments

Comments
 (0)