Machine learning in trading: theory, models, practice and algo-trading - page 738

 

Michael's foundation is fine. A good example for those who are looking for a place to start.

He has some good indicators, I think even from the stock exchange, which is already better than the standard mashka and rsishka. Then he does not just shove all this stuff into the neural network but uses different tools to evaluate the importance of indicators and select the most suitable ones for prediction. Then the neuronka is trained, but not a simple one, but one with built-in overfit control. And this entire strategy is tested in the tester using walk-forward.

SanSanych has already described a similar scheme at the beginning of this thread, finally someone decided to work with it. I think that many traders would like to put more indicators in their neuronics and not think.

I would also pre-process indicators through vtreat or caret. And a more adequate neuronka with kfold. And I would take more examples for training, I agree with the other critics.

 

I know, brothers!!! I'll answer you all at once... Everyone dances to the extent of his education, but our dispute will be resolved by the end result. The result, so to speak. I'm winning so far, because the deposit is growing. I know my approach is one of the correct one.

By the way, I'm just vtreat and select important predictors + tricks with shuffling.

This will be discussed in detail in the article about BOO.

 
Mihail Marchukajtes:

I know, brothers! I'll answer you all at once... Everyone dances to the extent of his education, but our dispute will be resolved by the end result. The result, so to speak. I'm winning so far, because the deposit is growing. I know my approach is one of the correct one.

By the way, it's vtreat and I'm selecting important predictors + tricks with shuffling.

This will be discussed in detail in the article about BO.

Above you wrote about mutual information. What does it have to do with it?

 
SanSanych Fomenko:

You wrote above about mutual information. What does it have to do with it?

The decisive criterion in choosing which model to use. It prepares the ditset. We train 10 times. We obtain 10 models with the same training results (as an example), where the number of training errors is the same for all models. Then we save the results of the models in the training section not in binary form, but in a dable. The result of polynomial calculation is initially considered in the dable......

Next, we count the VI of polynomials with respect to the output. Working models are those which have VI above the output entropy or more than 0.7 approximately....

Choose a model with high VI, but not higher than 0.95. Although I have not exactly checked it yet.... IMHO

 

As for increasing the training sample. Yes, it can be increased, but the quality of the network will decrease, which leads to a decrease in transactions. The network begins to work longer in time, but the number of "Don't know" states increases and it turns out that the network runs less often. With this approach it is necessary to train several models and run them in parallel.


Especially since the training period is now dictated to me by vtreat.R. Which with my input data offers me sets of inputs for different sample lengths. I choose the maximum learning period with the maximum number of input variables it chooses....And strangely enough, it spins from 20 to 50 samples. Apparently my inputs are like that. What can I do...

 

Mihail Marchukajtes:

...

The decisive criterion for choosing which model to use

...

I'll try to translate it into MO:

Although Mikhail's model is a classification model, it can return the result not only binary 0 and 1, but also the probability of belonging to a class (a fractional number from 0 to 1). Therefore, the model can be estimated not only with classification metrics, but also with regression metrics. I don't know why I chose mutual information, but it correlates well with MNC, so it's OK.

You can calculate it in R like this
library(entropy)
mi.plugin(rbind(target,prediction))
Here the smaller the result the better.

 
I had to make a custom neural network component, which beat TensorFlow's DNN:

Of course, but in statistics error (variation) is proportional to 1/sqrt(N) where N is the number of observations, here everything depends on the initial variation, the more it is, the more observations are needed to fit error into 5% quantile

"Look for someone who benefits from it" ((c) Lenin)

Only useless or even harmful tools are put and sold in public, for obvious reasons, this trivial truth should always be kept in mind, algotrading is not a webdesign or some other franchise, it's a direct functional theft of money from the market, everything that is open and easily accessible, much less advertised, a priori can not be useful in this context, junk or scam, such an aggressive and competitive environment we have.

Well, yes, the 5 percentile of winnings and an infinite abundance of publicly available tools is Klondike.

And I recently had to make a custom neural network component, which by three orders of magnitude surpassed a smart DNN from TensorFlow.

And most importantly, there was no complicated math or advanced technology, rather a simple, true original solution.

Against this background, Michael's commitment to the author's development Reshetov, personally, I am not at all surprised.

 
I can't do it:

That's what I mean, just that. If I take 40 samples through my infrastructure, I can't vouch for the result either, and 4000 points is very little. If of course we are talking about combat strategies, not article strategies.

I adhere to the following rule of thumb - an EA on new data will not work for more than 10% of bars trained. If I have trained on 4000 bars, it means that it will work on no more than 400. For example, if I want to trade a week on m15, I need at least 4800 (4*24*5*10, 10 weeks) bars of history for training.

Some day MT5 will introduce parallel threads in experts to be able to optimize the Expert Advisor in the background, then it will be possible to conduct interesting experiments with an extremely low number of bars and perhaps I will be able to do it with 40 bars. Although I probably won't be able to. Even if I have less than 1000 samples for training the result of neuronics skews in favor of trend and as soon as it changes - everything is bad.

 
Mihail Marchukajtes:

As for increasing the training sample. Yes, it can be increased, but the quality of the network will decrease, which leads to a decrease in transactions. The network begins to work longer in time, but the number of "Don't know" states increases and it turns out that the network runs less often. With this approach it is necessary to train several models and run them in parallel.


Especially since the training period is now dictated to me by vtreat.R. Which with my input data offers me sets of inputs for different sample lengths. I choose the maximum learning period with the maximum number of input variables it chooses....And strangely enough, it spins from 20 to 50 samples. Apparently my inputs are like that. What can I do...

Not the quality of the network will decrease but there will be less overfits and your model will appear in all its glory and poverty

My Reshetov's neuron optimizes in 3 months 1000+ trades on 3 competent features, and it works in zero on the real. And other neurons optimize even better. I don't know why you keep talking nonsense with your 100 trades, you are so stupid, you want to calm yourself down or to prove something to somebody?
 

hmm... good afternoon!)

and the word "curvafitter" is something neural?

Reason: