Machine learning in trading: theory, models, practice and algo-trading - page 3177

 
Aleksey Nikolayev #:

Yes

I.e. I need to generate a target binary for a sample, let's say, and see how often quantum segments will be found by my method for different predictors, and so 10 times?

If the number of quantum segments will be found about the same number on average as it is now for all predictors, then the method does not work, do I understand the thought correctly?

 
Wouldn't it be enough just to mix the column with the target column?
Tail and other parameters of the series will remain the same. I think that's a plus.
 
Maxim Dmitrievsky #:

There is a risk of getting bogged down in pointless arguments again. What is the difference between a randomly found set that works on oos and one that was invented through the hardest mental suffering, but also without fundamental justification? When the method of validation is the same. Rhetorical question.

What is the difference between a random search and a search with an element of randomness of choice? ))

The goal is to detect rule breaking, and to do that you need to know what they look like in the correct form. In other words, what statistical descriptive features can indicate that it is indeed a pattern. If you know them, then deviation from these rules would be a signal to stop the model.

And, assembling a model from such rules still seems to be a more stable solution than assembling pure randomness.

A couple of rules that give the main profit may get into the random, but then the market may forget about them for a while. With random it is more likely to panic and switch off the model at the beginning of a drawdown, but with the approach I suggest - waiting for the order of rules to change, i.e. a drawdown can be recognised as a normal phenomenon.

And the effect described above can be clearly seen on the gifs that I made, where in one case the sample test in plus, and in another case - although the rules are all taken from the sample train, but their (rules) appearance is not uniform in time.

The bonus of the approach, ideally, is to avoid exam sampling altogether.

An alternative could be deep model estimation - there is some work in this direction as well.

As a result, using only a random model:

1. We don't know why it works.

2. We don't know why it stopped working.

3. We don't know how to "fix" it.

4. We don't know if it's time to stop trading.

5. The lifetime of a random model will be shorter because it contains erroneous rules.

At randomisation, I personally get about the following - out of 100 models 30 of 30 are selected within half a year in the profitable zone about 10. I want at least 50% to 50%, then we can make partfiles.

 
Forester #:
Wouldn't it be enough just to mix the column with the target column?
Tail and other parameters of the series will remain the same. I think that's a plus.

Can you elaborate - I don't get it.

 
Aleksey Vyazmikin #:
The question was rhetorical
A random model and a model obtained by random brute force are two big differences.
You always confuse your statements, I'm used to it :)
After obtaining one of these models, the neighbourhood is always known, for more detailed analysis.
This is usually a pointless exercise, as it is easier to run a new search, but it is always a possibility.

If uncertainty scares you so much, there is an absolutely unambiguous search for absolutely unambiguous patterns in time series, on top of which some ts is added. To do this, you don't have to destroy common sense by reformatting the signs, but find the patterns on the original BP.
 
Maxim Dmitrievsky #:
The question was rhetorical
A random model and a model obtained by random brute force are two big differences.
You always confuse your statements, I'm used to it :)
After obtaining one of these models, the neighbourhood is always known, for more detailed analysis.
This is usually a pointless exercise, as it is easier to run a new search, but it is always a possibility.

If uncertainty scares you so much, there is an absolutely unambiguous search for absolutely unambiguous patterns in time series, on top of which some tc is docked. To do this, you don't have to destroy common sense by reformatting signs, but find patterns on the original BP.

You have no need to understand what I am explaining, and for that reason I have no desire to waste energy on re-explanation. Stay with your opinion about the identity of the result with different approaches.

 
Aleksey Vyazmikin #:

You have no need to understand what I am explaining, and for this reason I have no desire to waste energy on a repeated explanation. Stay with your opinion about the identity of the result in different approaches.

Behold, for if you are not given a pattern in the original series, the Hilbert path will not lead you to your cherished goal. Your endeavours will turn into devilishness, and you will find an ignominious slaughter instead of paradise.
 
Aleksey Vyazmikin #:

I.e. I need to generate a target binary for a sample, let's say, and see how often quantum segments will be found by my technique for different predictors, and so 10 times?

If the number of quantum segments will be found about the same number on average as it is now for all predictors, then the method does not work, do I understand the thought correctly?

Well yes, the point is to repeat your procedure many times on a large number of obviously meaningless problems. Then see how the application on specific real data looks relative to them - if it doesn't stand out much, then the method is bad. Usually formalise by counting some number on each application of the method and draw a sample, and then see if the number calculated on real data falls far into its tail - if yes, then ok.

 
Maxim Dmitrievsky #:
Behold, for if you are not given a pattern in the original series, the Hilbertian path will not lead you to your cherished goal. Your endeavours will turn into devilishness, and you will find an ignominious slaughter instead of paradise.

The problem of sectarians is the fear of testing their religious dogmas.

There are always many patterns - the question is the right choice.

 
Forester #:
Wouldn't it be enough just to mix the column with the target column?
Tail and other parameters of the series will remain the same. I think that's a plus.

It's not bad either. It's probably better to try both methods. If there are a lot of outliers in the traits (heavy tails), the results may be different, which may give additional information.

Reason: