Machine learning in trading: theory, models, practice and algo-trading - page 3494
You are missing trading opportunities:
- Free trading apps
- Over 8,000 signals for copying
- Economic news for exploring financial markets
Registration
Log in
You agree to website policy and terms of use
If you do not have an account, please register
I can fiddle with one sample so that it will pass the test on 5 more samples, but on the real 6th sample it will fail.
Can we see at least four samples?
I am thinking how to make an artificial sample for binary classification. Requirements - many predictors with noise and part of them with useful data, at least thousands of 20000 strings. The main thing is to know for sure that a solution can be found. Something like a benchmark to see how bad the algorithm is. I'm thinking of taking some function and marking its sections - above and below zero - respectively "1" and "0", any ideas?
Usually just generate multivariate random samples. If fully separable classes are needed, distributions with limited carriers are used - uniform, for example. If partial mixing of classes is needed, then Gaussian or their mixtures.
I want something meaningful, not random.
How do you see it other than the way it was in the video?
I want something meaningful, not random.
You want to reinvent the wheel, you have every right. I just described the standard practice.
I did not understand how the target markup is done in the proposed method, how the connection with predictors is established.
Why I want a meaningful something - because my goal is to reduce classification error through estimation of unstable quantum segments, and for that I want to look at the areas where error occurs - visualise them perhaps. To try different approach to estimation and to see the result is reliable, that for example on a significant predictor only 5 useful sites out of 10 were detected, I need to understand how they differ.
I didn't understand how the markup of the target is done in the proposed method, how the link to predictors is established.
Why I want something meaningful - because my task is to reduce classification error through estimation of unstable quantum segments, and for this purpose I want to look at areas where error occurs - visualise them, maybe. To try different approach to estimation and to see the result is reliable, that for example on a significant predictor only 5 useful sites out of 10 were detected, I need to understand how they differ.
Might be worth looking into data generation packages, something like this.
Thanks, might be useful, but it's not quite what I need. However, I've already got my own bicycle in my head.....
Well, in various ways, through mutual information, for example.
I'm not sure what to compare it to.