Discussing the article: "Neural networks made easy (Part 39): Go-Explore, a different approach to exploration"
Good afternoon Dimitri. Thank you for such a wonderful series of articles. I have tried all your Expert Advisors, but I have a problem with the latest ones.
The Expert Advisor from article 36 (the one with the largest neural network) passes the test in the tester, but the video card does not load during the test and the Expert Advisor does not try to trade. The balance graph does not change. There are no errors in the tester log. 1 kb files appear in the Common\Files folder.
Expert Advisors from articles 37, 38 are not tested at all. The test is started, but there is no progress. At the same time, the video card is loaded at 100%. And so on until the MT5 process is killed in the Manager. No files are created in Common\Files. There are no errors in the tester.
The Expert Advisor from this article Faza1 passes the test correctly without errors in the log, but the file GoExploer.bd is also created with the size of 1kb.
Can you please tell me where to dig? Other Expert Advisors from this series of articles (which are thrown on the chart) work normally and are considered as a video card. Video card RTX 3060 c 12Gb.
The Expert Advisor from this article Faza1 also passes the test correctly without errors in the log, but the GoExploer.bd file is also created with the size of 1kb.
On the demo he's specifically minus. It fills up at every opening of a new bar. I don't understand how he made a plus in the tester.
What is the training period? A short training period allows only to see if the model can learn. But not such experience is not enough to interpolate it to future states of the system.

- Free trading apps
- Over 8,000 signals for copying
- Economic news for exploring financial markets
You agree to website policy and terms of use
Check out the new article: Neural networks made easy (Part 39): Go-Explore, a different approach to exploration.
We continue studying the environment in reinforcement learning models. And in this article we will look at another algorithm – Go-Explore, which allows you to effectively explore the environment at the model training stage.
The main idea of Go-Explore is to remember and return to promising states. This is fundamental for effective operation when the number of rewards is limited. This idea is so flexible and broad that it can be implemented in a variety of ways.
Unlike most reinforcement learning algorithms, Go-Explore does not focus on directly solving the target problem, but rather on finding relevant states and actions in the state space that can lead to achieving the target state. To achieve this, the algorithm has two main phases: search and reuse.
The first phase is to go through all the states in the state space and record each state visited in a state "map". After this, the algorithm begins to study each visited state in more detail and to collect information about actions that can lead to other interesting states.
The second phase is to reuse previously learned states and actions to find new solutions. The algorithm stores the most successful trajectories and uses them to generate new states that can lead to even more successful solutions.
Author: Dmitriy Gizlyk