Dmitriy Gizlyk
Dmitriy Gizlyk
4.4 (50)
  • Information
12+ years
experience
0
products
0
demo versions
134
jobs
0
signals
0
subscribers
X
Professional programming of any complexity for MT4, MT5, C#.
Dmitriy Gizlyk
Published article Neural Networks in Trading: An Ensemble of Agents with Attention Mechanisms (MASAAT)
Neural Networks in Trading: An Ensemble of Agents with Attention Mechanisms (MASAAT)

We introduce the Multi-Agent Self-Adaptive Portfolio Optimization Framework (MASAAT), which combines attention mechanisms and time series analysis. MASAAT generates a set of agents that analyze price series and directional changes, enabling the identification of significant fluctuations in asset prices at different levels of detail.

Dmitriy Gizlyk
Published article Neural Networks in Trading: A Multi-Agent Self-Adaptive Model (Final Part)
Neural Networks in Trading: A Multi-Agent Self-Adaptive Model (Final Part)

In the previous article, we introduced the multi-agent self-adaptive framework MASA, which combines reinforcement learning approaches and self-adaptive strategies, providing a harmonious balance between profitability and risk in turbulent market conditions. We have built the functionality of individual agents within this framework. In this article, we will continue the work we started, bringing it to its logical conclusion.

Dmitriy Gizlyk
Published article Neural Networks in Trading: A Multi-Agent Self-Adaptive Model (MASA)
Neural Networks in Trading: A Multi-Agent Self-Adaptive Model (MASA)

I invite you to get acquainted with the Multi-Agent Self-Adaptive (MASA) framework, which combines reinforcement learning and adaptive strategies, providing a harmonious balance between profitability and risk management in turbulent market conditions.

Dmitriy Gizlyk
Published article Neural Networks in Trading: Parameter-Efficient Transformer with Segmented Attention (Final Part)
Neural Networks in Trading: Parameter-Efficient Transformer with Segmented Attention (Final Part)

In the previous work, we discussed the theoretical aspects of the PSformer framework, which includes two major innovations in the classical Transformer architecture: the Parameter Shared (PS) mechanism and attention to spatio-temporal segments (SegAtt). In this article, we continue the work we started on implementing the proposed approaches using MQL5.

Dmitriy Gizlyk
Published article Neural Networks in Trading: A Parameter-Efficient Transformer with Segmented Attention (PSformer)
Neural Networks in Trading: A Parameter-Efficient Transformer with Segmented Attention (PSformer)

This article introduces the new PSformer framework, which adapts the architecture of the vanilla Transformer to solving problems related to multivariate time series forecasting. The framework is based on two key innovations: the Parameter Sharing (PS) mechanism and the Segment Attention (SegAtt).

youwei_qing
youwei_qing 2025.04.21
I noticed that the feedForward method with the second input parameter isn't working at all. Could this be an issue? virtual bool feedForward(CNeuronBaseOCL *NeuronOCL); virtual bool feedForward(CNeuronBaseOCL *NeuronOCL, CBufferFloat *SecondInput) { return feedForward(NeuronOCL); }
Dmitriy Gizlyk
Published article Neural Networks in Trading: Enhancing Transformer Efficiency by Reducing Sharpness (Final Part)
Neural Networks in Trading: Enhancing Transformer Efficiency by Reducing Sharpness (Final Part)

SAMformer offers a solution to the key drawbacks of Transformer models in long-term time series forecasting, such as training complexity and poor generalization on small datasets. Its shallow architecture and sharpness-aware optimization help avoid suboptimal local minima. In this article, we will continue to implement approaches using MQL5 and evaluate their practical value.

Dmitriy Gizlyk
Published article Neural Networks in Trading: Enhancing Transformer Efficiency by Reducing Sharpness (SAMformer)
Neural Networks in Trading: Enhancing Transformer Efficiency by Reducing Sharpness (SAMformer)

Training Transformer models requires large amounts of data and is often difficult since the models are not good at generalizing to small datasets. The SAMformer framework helps solve this problem by avoiding poor local minima. This improves the efficiency of models even on limited training datasets.

Dmitriy Gizlyk
Published article Neural Networks in Trading: Optimizing the Transformer for Time Series Forecasting (LSEAttention)
Neural Networks in Trading: Optimizing the Transformer for Time Series Forecasting (LSEAttention)

The LSEAttention framework offers improvements to the Transformer architecture. It was designed specifically for long-term multivariate time series forecasting. The approaches proposed by the authors of the method can be applied to solve problems of entropy collapse and learning instability, which are often encountered with vanilla Transformer.

Dmitriy Gizlyk
Published article Neural Networks in Trading: Hyperbolic Latent Diffusion Model (Final Part)
Neural Networks in Trading: Hyperbolic Latent Diffusion Model (Final Part)

The use of anisotropic diffusion processes for encoding the initial data in a hyperbolic latent space, as proposed in the HypDIff framework, assists in preserving the topological features of the current market situation and improves the quality of its analysis. In the previous article, we started implementing the proposed approaches using MQL5. Today we will continue the work we started and will bring it to its logical conclusion.

Dmitriy Gizlyk
Published article Neural Networks in Trading: Hyperbolic Latent Diffusion Model (HypDiff)
Neural Networks in Trading: Hyperbolic Latent Diffusion Model (HypDiff)

The article considers methods of encoding initial data in hyperbolic latent space through anisotropic diffusion processes. This helps to more accurately preserve the topological characteristics of the current market situation and improves the quality of its analysis.

Dmitriy Gizlyk
Published article Neural Networks in Trading: Directional Diffusion Models (DDM)
Neural Networks in Trading: Directional Diffusion Models (DDM)

In this article, we discuss Directional Diffusion Models that exploit data-dependent anisotropic and directed noise in a forward diffusion process to capture meaningful graph representations.

Dmitriy Gizlyk
Published article Neural Networks in Trading: Node-Adaptive Graph Representation with NAFS
Neural Networks in Trading: Node-Adaptive Graph Representation with NAFS

We invite you to get acquainted with the NAFS (Node-Adaptive Feature Smoothing) method, which is a non-parametric approach to creating node representations that does not require parameter training. NAFS extracts features of each node given its neighbors and then adaptively combines these features to form a final representation.

Dmitriy Gizlyk
Published article Neural Networks in Trading: Contrastive Pattern Transformer (Final Part)
Neural Networks in Trading: Contrastive Pattern Transformer (Final Part)

In the previous last article within this series, we looked at the Atom-Motif Contrastive Transformer (AMCT) framework, which uses contrastive learning to discover key patterns at all levels, from basic elements to complex structures. In this article, we continue implementing AMCT approaches using MQL5.

Dmitriy Gizlyk
Published article Neural Networks in Trading: Contrastive Pattern Transformer
Neural Networks in Trading: Contrastive Pattern Transformer

The Contrastive Transformer is designed to analyze markets both at the level of individual candlesticks and based on entire patterns. This helps improve the quality of market trend modeling. Moreover, the use of contrastive learning to align representations of candlesticks and patterns fosters self-regulation and improves the accuracy of forecasts.

Dmitriy Gizlyk
Published article Neural Networks in Trading: Market Analysis Using a Pattern Transformer
Neural Networks in Trading: Market Analysis Using a Pattern Transformer

When we use models to analyze the market situation, we mainly focus on the candlestick. However, it has long been known that candlestick patterns can help in predicting future price movements. In this article, we will get acquainted with a method that allows us to integrate both of these approaches.

Dmitriy Gizlyk
Published article Neural Networks in Trading: Transformer with Relative Encoding
Neural Networks in Trading: Transformer with Relative Encoding

Self-supervised learning can be an effective way to analyze large amounts of unlabeled data. The efficiency is provided by the adaptation of models to the specific features of financial markets, which helps improve the effectiveness of traditional methods. This article introduces an alternative attention mechanism that takes into account the relative dependencies and relationships between inputs.

Dmitriy Gizlyk
Published article Neural Networks in Trading: Controlled Segmentation (Final Part)
Neural Networks in Trading: Controlled Segmentation (Final Part)

We continue the work started in the previous article on building the RefMask3D framework using MQL5. This framework is designed to comprehensively study multimodal interaction and feature analysis in a point cloud, followed by target object identification based on a description provided in natural language.

Dmitriy Gizlyk
Published article Neural Networks in Trading: Controlled Segmentation
Neural Networks in Trading: Controlled Segmentation

In this article. we will discuss a method of complex multimodal interaction analysis and feature understanding.

Dmitriy Gizlyk
Published article Neural Networks in Trading: Generalized 3D Referring Expression Segmentation
Neural Networks in Trading: Generalized 3D Referring Expression Segmentation

While analyzing the market situation, we divide it into separate segments, identifying key trends. However, traditional analysis methods often focus on one aspect and thus limit the proper perception. In this article, we will learn about a method that enables the selection of multiple objects to ensure a more comprehensive and multi-layered understanding of the situation.

Dmitriy Gizlyk
Published article Neural Networks in Trading: Mask-Attention-Free Approach to Price Movement Forecasting
Neural Networks in Trading: Mask-Attention-Free Approach to Price Movement Forecasting

In this article, we will discuss the Mask-Attention-Free Transformer (MAFT) method and its application in the field of trading. Unlike traditional Transformers that require data masking when processing sequences, MAFT optimizes the attention process by eliminating the need for masking, significantly improving computational efficiency.