Machine learning in trading: theory, models, practice and algo-trading - page 2749

You are missing trading opportunities:
- Free trading apps
- Over 8,000 signals for copying
- Economic news for exploring financial markets
Registration
Log in
You agree to website policy and terms of use
If you do not have an account, please register
added in the previous post
#299
someone deleted my posts where I posted the vid... i don't know how that can happen.
someone deleted my posts where I posted the vid... i don't know how that can happen.
Well, well, well, you're showing off a lot, I guess.
Here's one on wood modelling.
#4469
#4467
About the catbusters search? 😄 who started the thread
Well, well, well, you must be showing off a lot.
here's the wood modelling
#4469
#4467
About the catbusters search? 😄 who started the thread
Look, wood is not an argument at all... 95 per cent of the MoD tutorials start with wood, it's the first thing a beginner gets to know, you're not in charge of it.
Look, wood is not an argument at all... 95% of the MoD tutorials start with wood, that's what a beginner learns first, it's not your fault.
It's been mostly NS here before, I wondered why not trees. I made tests, it turned out that they are no worse. I haven't seen such information here before, otherwise I wouldn't have asked this question.
Then I switched over to bousting, found the katbust lib and started replicating it here
It was mostly NSs that were discussed here before, I wondered why not trees. I made tests and it turned out that they are not worse. I haven't seen such info here before, otherwise I wouldn't have asked this question.
and trees, by the way, can serve as a worthy alternative to clustering and fs - "2 in one".
Then I switched to bousting, found the katbust lib and started replicating it here
and trees, by the way, can serve as a worthy alternative to clustering and fs.
on the algorithm: it gives another 2nd derivative mathematically (or residuals averaging - statistically) -- BUT how does it help in training and in what cases?... besides the standard clichés in commercials "catboost will give better and more accurate results".... because not everywhere point precision is important, sometimes the generative ability of the model may be more important?There are also wooden models for causal inference, haven't had time to figure it out yet
Boosting reduces bias and variance, whereas Forest only variance I think. It's about proven advantages, you can google it. And the library itself is developed, it is convenient to work with it.
It is not quite clear about generative ones, maybe sometimes they are more important. But NS generative ones do not work well at Forex, if we are talking about generating synthetic data.
if we're talking about generating synthetic data.
no - it's about generalising... yes, wrong, express it.... sorry.
I think to distinguish risk-on/risk-off environment -- still thinking about how to generalise this division... all in my own thoughts (on the forum by accident)...
thanks for the reply!
no - it's about generalising..... yes, wrong, express.... sorry
I think to distinguish risk-on/risk-off environment -- still thinking about how to generalise this division... all in my own thoughts (on the forum by accident)...
thanks for the reply!
try it, catbusta has a bunch of different features, I like it.
There is an early stop based on error on validation sample, pre-training. Generalisation is not worse than in NS, which in addition have to write their own functions to stop learning.
and it learns quickly, you don't have to wait for hours.
I'm just wondering if the MO can simulate primitive functions from the JA
there is a matrix, each row of the matrix is a training example.
we need to find the maximum of each row, the sample size is 20k rows.
solving the problem through regression
rounded for clarity
pretty good, only a few errors on the test out of 50 new lines
But the data in matrix X is very simple, only 5 unique values from 1 to 5 and only 5 columns, but already errors.
Although I think if we make classification, there would be no errors, by the way you can check it.
Well, yes, it is, but if we are looking for the maximum in the data then classification is not suitable because the scatter of values can be huge....
So let's go back to regression and complicate the data.
we get this result.
In principle it is not bad, but the usual function mach() will do it better and can replace all this model....
By the way, I wonder how other models will work, if they will be able to create the mach() function without an error
I remember, but it's a fitting error, I mean, trace sampling...
It's the error on the next 300 bars. Predictors were generated on each bar, then filtered, the model was trained and the next bar was predicted.