Dmitriy Gizlyk
Dmitriy Gizlyk
4.4 (49)
  • Bilgiler
10+ yıl
deneyim
0
ürünler
0
demo sürümleri
134
işler
0
sinyaller
0
aboneler
Professional writing programs of any complexity for MT4, MT5, C#.
Dmitriy Gizlyk
"Neural networks made easy (Part 74): Trajectory prediction with adaptation" makalesini yayınladı
Neural networks made easy (Part 74): Trajectory prediction with adaptation

This article introduces a fairly effective method of multi-agent trajectory forecasting, which is able to adapt to various environmental conditions.

2
Dmitriy Gizlyk
"Neural networks made easy (Part 73): AutoBots for predicting price movements" makalesini yayınladı
Neural networks made easy (Part 73): AutoBots for predicting price movements

We continue to discuss algorithms for training trajectory prediction models. In this article, we will get acquainted with a method called "AutoBots".

1
Dmitriy Gizlyk
"Neural networks made easy (Part 72): Trajectory prediction in noisy environments" makalesini yayınladı
Neural networks made easy (Part 72): Trajectory prediction in noisy environments

The quality of future state predictions plays an important role in the Goal-Conditioned Predictive Coding method, which we discussed in the previous article. In this article I want to introduce you to an algorithm that can significantly improve the prediction quality in stochastic environments, such as financial markets.

2
Dmitriy Gizlyk
"Neural networks made easy (Part 71): Goal-Conditioned Predictive Coding (GCPC)" makalesini yayınladı
Neural networks made easy (Part 71): Goal-Conditioned Predictive Coding (GCPC)

In previous articles, we discussed the Decision Transformer method and several algorithms derived from it. We experimented with different goal setting methods. During the experiments, we worked with various ways of setting goals. However, the model's study of the earlier passed trajectory always remained outside our attention. In this article. I want to introduce you to a method that fills this gap.

3
Dmitriy Gizlyk
"Neural networks made easy (Part 70): Closed-Form Policy Improvement Operators (CFPI)" makalesini yayınladı
Neural networks made easy (Part 70): Closed-Form Policy Improvement Operators (CFPI)

In this article, we will get acquainted with an algorithm that uses closed-form policy improvement operators to optimize Agent actions in offline mode.

2
Dmitriy Gizlyk
"Neural networks made easy (Part 69): Density-based support constraint for the behavioral policy (SPOT)" makalesini yayınladı
Neural networks made easy (Part 69): Density-based support constraint for the behavioral policy (SPOT)

In offline learning, we use a fixed dataset, which limits the coverage of environmental diversity. During the learning process, our Agent can generate actions beyond this dataset. If there is no feedback from the environment, how can we be sure that the assessments of such actions are correct? Maintaining the Agent's policy within the training dataset becomes an important aspect to ensure the reliability of training. This is what we will talk about in this article.

2
JimReaper
JimReaper 2023.12.22
Hi Dmitriy, seems like the article is incomplete.
Dmitriy Gizlyk
"Neural networks made easy (Part 68): Offline Preference-guided Policy Optimization" makalesini yayınladı
Neural networks made easy (Part 68): Offline Preference-guided Policy Optimization

Since the first articles devoted to reinforcement learning, we have in one way or another touched upon 2 problems: exploring the environment and determining the reward function. Recent articles have been devoted to the problem of exploration in offline learning. In this article, I would like to introduce you to an algorithm whose authors completely eliminated the reward function.

2
Dmitriy Gizlyk
"Neural networks made easy (Part 67): Using past experience to solve new tasks" makalesini yayınladı
Neural networks made easy (Part 67): Using past experience to solve new tasks

In this article, we continue discussing methods for collecting data into a training set. Obviously, the learning process requires constant interaction with the environment. However, situations can be different.

4
JimReaper
JimReaper 2023.12.09
THIS IS GENIUS WORK Dmitriy! I Love this!
Dmitriy Gizlyk
"Neural networks made easy (Part 66): Exploration problems in offline learning" makalesini yayınladı
Neural networks made easy (Part 66): Exploration problems in offline learning

Models are trained offline using data from a prepared training dataset. While providing certain advantages, its negative side is that information about the environment is greatly compressed to the size of the training dataset. Which, in turn, limits the possibilities of exploration. In this article, we will consider a method that enables the filling of a training dataset with the most diverse data possible.

3
JimReaper
JimReaper 2023.12.05
You are the best! Thank you so much for your research. Beautifully done.!
Dmitriy Gizlyk
"Neural networks made easy (Part 65): Distance Weighted Supervised Learning (DWSL)" makalesini yayınladı
Neural networks made easy (Part 65): Distance Weighted Supervised Learning (DWSL)

In this article, we will get acquainted with an interesting algorithm that is built at the intersection of supervised and reinforcement learning methods.

2
Dmitriy Gizlyk
"Neural networks made easy (Part 64): ConserWeightive Behavioral Cloning (CWBC) method" makalesini yayınladı
Neural networks made easy (Part 64): ConserWeightive Behavioral Cloning (CWBC) method

As a result of tests performed in previous articles, we came to the conclusion that the optimality of the trained strategy largely depends on the training set used. In this article, we will get acquainted with a fairly simple yet effective method for selecting trajectories to train models.

1
Dmitriy Gizlyk
"Neural networks made easy (Part 63): Unsupervised Pretraining for Decision Transformer (PDT)" makalesini yayınladı
Neural networks made easy (Part 63): Unsupervised Pretraining for Decision Transformer (PDT)

We continue to discuss the family of Decision Transformer methods. From previous article, we have already noticed that training the transformer underlying the architecture of these methods is a rather complex task and requires a large labeled dataset for training. In this article we will look at an algorithm for using unlabeled trajectories for preliminary model training.

1
Dmitriy Gizlyk
"Neural networks made easy (Part 62): Using Decision Transformer in hierarchical models" makalesini yayınladı
Neural networks made easy (Part 62): Using Decision Transformer in hierarchical models

In recent articles, we have seen several options for using the Decision Transformer method. The method allows analyzing not only the current state, but also the trajectory of previous states and actions performed in them. In this article, we will focus on using this method in hierarchical models.

1
Dmitriy Gizlyk
"Neural networks made easy (Part 61): Optimism issue in offline reinforcement learning" makalesini yayınladı
Neural networks made easy (Part 61): Optimism issue in offline reinforcement learning

During the offline learning, we optimize the Agent's policy based on the training sample data. The resulting strategy gives the Agent confidence in its actions. However, such optimism is not always justified and can cause increased risks during the model operation. Today we will look at one of the methods to reduce these risks.

1
Dmitriy Gizlyk
"Neural networks made easy (Part 60): Online Decision Transformer (ODT)" makalesini yayınladı
Neural networks made easy (Part 60): Online Decision Transformer (ODT)

The last two articles were devoted to the Decision Transformer method, which models action sequences in the context of an autoregressive model of desired rewards. In this article, we will look at another optimization algorithm for this method.

1
Dmitriy Gizlyk
"Neural networks are easy (Part 59): Dichotomy of Control (DoC)" makalesini yayınladı
Neural networks are easy (Part 59): Dichotomy of Control (DoC)

In the previous article, we got acquainted with the Decision Transformer. But the complex stochastic environment of the foreign exchange market did not allow us to fully implement the potential of the presented method. In this article, I will introduce an algorithm that is aimed at improving the performance of algorithms in stochastic environments.

2
Dmitriy Gizlyk
"Neural networks made easy (Part 58): Decision Transformer (DT)" makalesini yayınladı
Neural networks made easy (Part 58): Decision Transformer (DT)

We continue to explore reinforcement learning methods. In this article, I will focus on a slightly different algorithm that considers the Agent’s policy in the paradigm of constructing a sequence of actions.

6
Yao Wei Lai
Yao Wei Lai 2023.10.11
I greatly admire your article series "Neural Networks Make It Easy", but after reading it for a long time, I still don't understand how to generate models. Could you please send me the models used in each article? I would like to replicate your test to further learn relevant knowledge. Thank you!
Dmitriy Gizlyk
"Neural networks made easy (Part 57): Stochastic Marginal Actor-Critic (SMAC)" makalesini yayınladı
Neural networks made easy (Part 57): Stochastic Marginal Actor-Critic (SMAC)

Here I will consider the fairly new Stochastic Marginal Actor-Critic (SMAC) algorithm, which allows building latent variable policies within the framework of entropy maximization.

5
Dmitriy Gizlyk
"Neural networks made easy (Part 56): Using nuclear norm to drive research" makalesini yayınladı
Neural networks made easy (Part 56): Using nuclear norm to drive research

The study of the environment in reinforcement learning is a pressing problem. We have already looked at some approaches previously. In this article, we will have a look at yet another method based on maximizing the nuclear norm. It allows agents to identify environmental states with a high degree of novelty and diversity.

3
Dmitriy Gizlyk
"Neural networks made easy (Part 55): Contrastive intrinsic control (CIC)" makalesini yayınladı
Neural networks made easy (Part 55): Contrastive intrinsic control (CIC)

Contrastive training is an unsupervised method of training representation. Its goal is to train a model to highlight similarities and differences in data sets. In this article, we will talk about using contrastive training approaches to explore different Actor skills.

6