Skip to content

Commit 8ebae1f

Browse files
Update README.md
1 parent 7287d69 commit 8ebae1f

File tree

1 file changed

+9
-0
lines changed

1 file changed

+9
-0
lines changed

README.md

Lines changed: 9 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -10,6 +10,8 @@ Valentyn N Sichkar. Reinforcement Learning Algorithms for global path planning /
1010

1111
* The research results for Neural Network Knowledge Based system for the tasks of collision avoidance is put in separate repository and is available here: https://github.com/sichkar-valentyn/Matlab_implementation_of_Neural_Networks
1212

13+
* The study of Semantic Web languages OWL and RDF for Knowledge representation of Alarm-Warning System is put in separate repository and is available here: https://github.com/sichkar-valentyn/Knowledge_Base_Represented_by_Semantic_Web_Language
14+
1315
* The study of Neural Networks for Computer Vision in autonomous vehicles and robotics is put in separate repository and is available here: https://github.com/sichkar-valentyn/Neural_Networks_for_Computer_Vision
1416

1517
## Description
@@ -77,12 +79,15 @@ Environment-1 with mobile robot, goal and obstacles
7779

7880
<img src="images/Environment-1.gif" alt="RL_Q-Learning_E-1" width=362 height=391> <img src="images/Environment-1.png" alt="RL_Q-Learning_E-1" width=362 height=391>
7981

82+
<br/>
8083

8184
### <a name="Q-learning algorithm resulted chart for the environment-1">Q-learning algorithm resulted chart for the environment-1</a>
8285
Represents number of episodes via number of steps and number of episodes via cost for each episode
8386

8487
![RL_Q-Learning_C-1](images/Charts-1.png)
8588

89+
<br/>
90+
8691
### <a name="Final Q-table with values from the final shortest route for environment-1">Final Q-table with values from the final shortest route for environment-1</a>
8792
![RL_Q-Learning_T-1](images/Q-Table-E-1.png)
8893
<br/>Looking at the values of the table we can see the decision for the next action made by agent (mobile robot). The sequence of final actions to reach the goal after the Q-table is filled with knowledge is the following: *down-right-down-down-down-right-down-right-down-right-down-down-right-right-up-up.*
@@ -95,11 +100,15 @@ Bigger environment-2 with more obstacles
95100

96101
![RL_Q-Learning_E-2](images/Environment-2.png)
97102

103+
<br/>
104+
98105
### <a name="Q-learning algorithm resulted chart for the environment-2">Q-learning algorithm resulted chart for the environment-2</a>
99106
Represents number of episodes via number of steps and number of episodes via cost for each episode
100107

101108
![RL_Q-Learning_C-2](images/Charts-2.png)
102109

110+
<br/>
111+
103112
### <a name="Final Q-table with values from the final shortest route for environment-1">Final Q-table with values from the final shortest route for environment-1</a>
104113
![RL_Q-Learning_T-2](images/Q-Table-E-2.png)
105114

0 commit comments

Comments
 (0)