Dmitriy Gizlyk
Dmitriy Gizlyk
4.4 (50)
  • Informations
12+ années
expérience
0
produits
0
versions de démo
134
offres d’emploi
0
signaux
0
les abonnés
Professional programming of any complexity for MT4, MT5, C#.
Dmitriy Gizlyk
Article publié Neural Networks Made Easy (Part 87): Time Series Patching
Neural Networks Made Easy (Part 87): Time Series Patching

Forecasting plays an important role in time series analysis. In the new article, we will talk about the benefits of time series patching.

2
Dmitriy Gizlyk
Article publié Neural Networks Made Easy (Part 86): U-Shaped Transformer
Neural Networks Made Easy (Part 86): U-Shaped Transformer

We continue to study timeseries forecasting algorithms. In this article, we will discuss another method: the U-shaped Transformer.

3
Dmitriy Gizlyk
Article publié Neural Networks Made Easy (Part 85): Multivariate Time Series Forecasting
Neural Networks Made Easy (Part 85): Multivariate Time Series Forecasting

In this article, I would like to introduce you to a new complex timeseries forecasting method, which harmoniously combines the advantages of linear models and transformers.

2
Dmitriy Gizlyk
Article publié Neural Networks Made Easy (Part 84): Reversible Normalization (RevIN)
Neural Networks Made Easy (Part 84): Reversible Normalization (RevIN)

We already know that pre-processing of the input data plays a major role in the stability of model training. To process "raw" input data online, we often use a batch normalization layer. But sometimes we need a reverse procedure. In this article, we discuss one of the possible approaches to solving this problem.

4
Dmitriy Gizlyk
Article publié Neural Networks Made Easy (Part 83): The "Conformer" Spatio-Temporal Continuous Attention Transformer Algorithm
Neural Networks Made Easy (Part 83): The "Conformer" Spatio-Temporal Continuous Attention Transformer Algorithm

This article introduces the Conformer algorithm originally developed for the purpose of weather forecasting, which in terms of variability and capriciousness can be compared to financial markets. Conformer is a complex method. It combines the advantages of attention models and ordinary differential equations.

6
Look Mode
Look Mode 2024.03.30
Здравствуйте, как эти файлы попробовать (тестировать) из файлы Comformer?
Dmitriy Gizlyk
Article publié Neural networks made easy (Part 82): Ordinary Differential Equation models (NeuralODE)
Neural networks made easy (Part 82): Ordinary Differential Equation models (NeuralODE)

In this article, we will discuss another type of models that are aimed at studying the dynamics of the environmental state.

5
Dmitriy Gizlyk
Article publié Neural Networks Made Easy (Part 81): Context-Guided Motion Analysis (CCMR)
Neural Networks Made Easy (Part 81): Context-Guided Motion Analysis (CCMR)

In previous works, we always assessed the current state of the environment. At the same time, the dynamics of changes in indicators always remained "behind the scenes". In this article I want to introduce you to an algorithm that allows you to evaluate the direct change in data between 2 successive environmental states.

2
Dmitriy Gizlyk
Article publié Neural networks made easy (Part 80): Graph Transformer Generative Adversarial Model (GTGAN)
Neural networks made easy (Part 80): Graph Transformer Generative Adversarial Model (GTGAN)

In this article, I will get acquainted with the GTGAN algorithm, which was introduced in January 2024 to solve complex problems of generation architectural layouts with graph constraints.

3
Dmitriy Gizlyk
Article publié Neural networks made easy (Part 79): Feature Aggregated Queries (FAQ) in the context of state
Neural networks made easy (Part 79): Feature Aggregated Queries (FAQ) in the context of state

In the previous article, we got acquainted with one of the methods for detecting objects in an image. However, processing a static image is somewhat different from working with dynamic time series, such as the dynamics of the prices we analyze. In this article, we will consider the method of detecting objects in video, which is somewhat closer to the problem we are solving.

3
Dmitriy Gizlyk
Article publié Neural networks made easy (Part 78): Decoder-free Object Detector with Transformer (DFFT)
Neural networks made easy (Part 78): Decoder-free Object Detector with Transformer (DFFT)

In this article, I propose to look at the issue of building a trading strategy from a different angle. We will not predict future price movements, but will try to build a trading system based on the analysis of historical data.

4
Dmitriy Gizlyk
Article publié Neural networks made easy (Part 77): Cross-Covariance Transformer (XCiT)
Neural networks made easy (Part 77): Cross-Covariance Transformer (XCiT)

In our models, we often use various attention algorithms. And, probably, most often we use Transformers. Their main disadvantage is the resource requirement. In this article, we will consider a new algorithm that can help reduce computing costs without losing quality.

2
Dmitriy Gizlyk
Article publié Neural networks made easy (Part 76): Exploring diverse interaction patterns with Multi-future Transformer
Neural networks made easy (Part 76): Exploring diverse interaction patterns with Multi-future Transformer

This article continues the topic of predicting the upcoming price movement. I invite you to get acquainted with the Multi-future Transformer architecture. Its main idea is to decompose the multimodal distribution of the future into several unimodal distributions, which allows you to effectively simulate various models of interaction between agents on the scene.

1
Dmitriy Gizlyk
Article publié Neural networks made easy (Part 75): Improving the performance of trajectory prediction models
Neural networks made easy (Part 75): Improving the performance of trajectory prediction models

The models we create are becoming larger and more complex. This increases the costs of not only their training as well as operation. However, the time required to make a decision is often critical. In this regard, let us consider methods for optimizing model performance without loss of quality.

2
Dmitriy Gizlyk
Article publié Neural networks made easy (Part 74): Trajectory prediction with adaptation
Neural networks made easy (Part 74): Trajectory prediction with adaptation

This article introduces a fairly effective method of multi-agent trajectory forecasting, which is able to adapt to various environmental conditions.

2
Dmitriy Gizlyk
Article publié Neural networks made easy (Part 73): AutoBots for predicting price movements
Neural networks made easy (Part 73): AutoBots for predicting price movements

We continue to discuss algorithms for training trajectory prediction models. In this article, we will get acquainted with a method called "AutoBots".

1
Dmitriy Gizlyk
Article publié Neural networks made easy (Part 72): Trajectory prediction in noisy environments
Neural networks made easy (Part 72): Trajectory prediction in noisy environments

The quality of future state predictions plays an important role in the Goal-Conditioned Predictive Coding method, which we discussed in the previous article. In this article I want to introduce you to an algorithm that can significantly improve the prediction quality in stochastic environments, such as financial markets.

2
Dmitriy Gizlyk
Article publié Neural networks made easy (Part 71): Goal-Conditioned Predictive Coding (GCPC)
Neural networks made easy (Part 71): Goal-Conditioned Predictive Coding (GCPC)

In previous articles, we discussed the Decision Transformer method and several algorithms derived from it. We experimented with different goal setting methods. During the experiments, we worked with various ways of setting goals. However, the model's study of the earlier passed trajectory always remained outside our attention. In this article. I want to introduce you to a method that fills this gap.

3
Dmitriy Gizlyk
Article publié Neural networks made easy (Part 70): Closed-Form Policy Improvement Operators (CFPI)
Neural networks made easy (Part 70): Closed-Form Policy Improvement Operators (CFPI)

In this article, we will get acquainted with an algorithm that uses closed-form policy improvement operators to optimize Agent actions in offline mode.

2
Dmitriy Gizlyk
Article publié Neural networks made easy (Part 69): Density-based support constraint for the behavioral policy (SPOT)
Neural networks made easy (Part 69): Density-based support constraint for the behavioral policy (SPOT)

In offline learning, we use a fixed dataset, which limits the coverage of environmental diversity. During the learning process, our Agent can generate actions beyond this dataset. If there is no feedback from the environment, how can we be sure that the assessments of such actions are correct? Maintaining the Agent's policy within the training dataset becomes an important aspect to ensure the reliability of training. This is what we will talk about in this article.

2
JimReaper
JimReaper 2023.12.22
Hi Dmitriy, seems like the article is incomplete.
Dmitriy Gizlyk
Article publié Neural networks made easy (Part 68): Offline Preference-guided Policy Optimization
Neural networks made easy (Part 68): Offline Preference-guided Policy Optimization

Since the first articles devoted to reinforcement learning, we have in one way or another touched upon 2 problems: exploring the environment and determining the reward function. Recent articles have been devoted to the problem of exploration in offline learning. In this article, I would like to introduce you to an algorithm whose authors completely eliminated the reward function.

2