You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+9Lines changed: 9 additions & 0 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -10,6 +10,8 @@ Valentyn N Sichkar. Reinforcement Learning Algorithms for global path planning /
10
10
11
11
* The research results for Neural Network Knowledge Based system for the tasks of collision avoidance is put in separate repository and is available here: https://github.com/sichkar-valentyn/Matlab_implementation_of_Neural_Networks
12
12
13
+
* The study of Semantic Web languages OWL and RDF for Knowledge representation of Alarm-Warning System is put in separate repository and is available here: https://github.com/sichkar-valentyn/Knowledge_Base_Represented_by_Semantic_Web_Language
14
+
13
15
* The study of Neural Networks for Computer Vision in autonomous vehicles and robotics is put in separate repository and is available here: https://github.com/sichkar-valentyn/Neural_Networks_for_Computer_Vision
14
16
15
17
## Description
@@ -77,12 +79,15 @@ Environment-1 with mobile robot, goal and obstacles
### <aname="Q-learning algorithm resulted chart for the environment-1">Q-learning algorithm resulted chart for the environment-1</a>
82
85
Represents number of episodes via number of steps and number of episodes via cost for each episode
83
86
84
87

85
88
89
+
<br/>
90
+
86
91
### <aname="Final Q-table with values from the final shortest route for environment-1">Final Q-table with values from the final shortest route for environment-1</a>
87
92

88
93
<br/>Looking at the values of the table we can see the decision for the next action made by agent (mobile robot). The sequence of final actions to reach the goal after the Q-table is filled with knowledge is the following: *down-right-down-down-down-right-down-right-down-right-down-down-right-right-up-up.*
@@ -95,11 +100,15 @@ Bigger environment-2 with more obstacles
95
100
96
101

97
102
103
+
<br/>
104
+
98
105
### <aname="Q-learning algorithm resulted chart for the environment-2">Q-learning algorithm resulted chart for the environment-2</a>
99
106
Represents number of episodes via number of steps and number of episodes via cost for each episode
100
107
101
108

102
109
110
+
<br/>
111
+
103
112
### <aname="Final Q-table with values from the final shortest route for environment-1">Final Q-table with values from the final shortest route for environment-1</a>
0 commit comments