Machine learning in trading: theory, models, practice and algo-trading - page 1704
You are missing trading opportunities:
- Free trading apps
- Over 8,000 signals for copying
- Economic news for exploring financial markets
Registration
Log in
You agree to website policy and terms of use
If you do not have an account, please register
A mistake is not just a mistake. A small error can have a big impact.
And NS is not asked to capture repetitive data. It is being asked to identify hidden patterns to get the right result in the absence of repetitive data. Understanding Generalization. When we have a finite domain of data, but only have 50% of that data. The network learns and identifying a pattern can build the rest of the data that it hasn't seen. It's like reconstructing old video footage with missing pixels, which the network reconstructs on its own.
Well, that's a statistical approach! For example, I have 20 pictures of different triangles. On the 21st one, the triangle is missing the hypotenuse. I can easily determine that this is an undrawn triangle based on the statistics I've gathered - every photo has a triangle in it. That is, an error in the drawing of the triangle is not critical for me, because I already have statistics in general. Doesn't this principle work with NS?
Well approximately, if you don't count that in 21 pictures you have just a corner drawn and the triangle is out of the question there...
To understand the essence and theoretical basis of neural networks it is necessary to know Kolmogorov, Arnold and Hecht-Nielson theory.
This knowledge is not particularly necessary for practice, but a general understanding of it would not hurt.
I gave you a link to view splits from JOT data. That's where the full model is unloaded to a file. Then the splits are read from it.
Are you sure you can unload asymmetric trees already?
In boosting, by definition, all trees are important. Each successive tree refines all the previous ones. If you throw out one tree in the middle, all the trees after that will work with wrong data - they have to be retrained, without taking into account the thrown out tree. The first tree will very closely replicate the discarded one.
No, that's not exactly true. When the model is ready there's a bunch of tree leaves sitting in it, giving a probability near 0.5 - which is garbage in essence.
Yes. The individual leaves in the boosting are incomplete because they are augmented by the responses of leaves from other refining trees. And only the aggregate of the answers of, say, 100 trees gives the correct answer.
Trying to get something reliable from one sheet of the boosting model is impossible.
In boosting, all 100 responses from 100 trees are summed up, each giving for example 0.01 in total = 1. The value of 1 leaf = 0.01 - what do you want to get from it? There is nothing in it. Only the sum of 100 leaves will give the correct answer.
In fact, there the 1st tree is strong and gives for example 0.7, the rest bring the sum closer to 1. If only the leaves of the first tree are considered separately, but I think they are weaker than any tree from the random forest, due to less depth.
In a random forest the average is found, e.g. every leaf of 100 trees = 1, the average also = 1. In it the foxes are complete, but with random deviations. But a crowd of 100 answers, gives the average as a fairly accurate answer.
It is possible to get useful from a single tree, but it is rare.
Even the first tree will not be the best.
Boosting is fast, but you have to carefully select models, and that requires a large sample, which is a problem.
To understand the essence and theoretical basis of neural networks it is necessary to know Kolmogorov, Arnold and Hecht-Nielson theory.
For practice, this knowledge is not particularly necessary, but a general understanding of it would not hurt.
Spc.
Lack of rigorous theory related to the mentioned above models of neural networks does not prevent from researching possibilities of their application to practical tasks.
Are you sure you can unload asymmetric trees already?
Tried an example by adding grow_policy='Lossguide'.
Here's a piece of the model containing the splits:
There are splits with Depthwise, too. But this is in Python, I haven't seen a way to unload a model in R. But I think you can save the model in R in an internal format, open it in Python and unload it to JSON. And from it all you need to use
Colleagues. This topic is interesting, but I do not understand the difference between machine learning and simple optimization of parameters in the tester. The purpose will be approximately the same - making decision about market entry (exit) when input (exit) parameters coincide. Both methods stop working when the market strongly changes. It is necessary to train and optimize anew. In brief and without philosophy.
The difference is in the flexibility of the tool. The neural network can adapt to any series. Look at the polynomial below. The digit capacity of coefficients as well as their number can be absolutely different.
Now imagine the number of variants. No multi-parameter Expert Advisor can boast of such flexibility. They are also said to have a generalizing capability, but that's not exactly true :-)