Machine learning in trading: theory, models, practice and algo-trading - page 31
You are missing trading opportunities:
- Free trading apps
- Over 8,000 signals for copying
- Economic news for exploring financial markets
Registration
Log in
You agree to website policy and terms of use
If you do not have an account, please register
Yuri, the first sample on your data:
Two different sets of parameter values for training. It is noteworthy that the AUC is below the plinth on crossvalidation.
All in all, 51.5% accuracy on the test is the best we got.
I don't even know how you get about 60%.
You should throw out that set of predictors.
If we just take every single increment, a few oscillators more than 100 predictors with more than 5000 observations, i.e. H1, then from such a set we can choose 10-15 predictors, which not only have less than 40% prediction error, but most importantly they will give us a REAL model.
We should throw out this set of predictors.
If we stupidly take the increments of everything, a few oscillators over 100 predictors with more than 5000 observations, i.e. H1, then from such a set of 10-15 predictors can be selected, which not only give the prediction error less than 40%, but most importantly, they give a NOT REBUILDED model.
In general, I did not get better than 51.5% classification accuracy. Accordingly, the rest of the metrics will also be close to random guessing.
The balance of responses on the test is almost perfectly 50/50.
Yuri, I look forward to your revelations.
In general, I did not get a better than 51.5% classification accuracy. Correspondingly, the rest of the metrics will also be close to random guessing.
The balance of answers on the test is almost perfectly 50/50.
Yuri, I am waiting for your revelations.
I am not hiding anything. For the old version the results of which I have already given above, all information is in the public domain:
Description of the method of building a binary classifier: https://sites.google.com/site/libvmr/
Java source code with comments: https://sourceforge.net/p/libvmr/code/HEAD/tree/trunk/
Builds: https://sourceforge.net/projects/libvmr/files/
Yuri, thank you.
If the set is linearly separable, then the number of potential separating hyperplanes is infinite. In that case, it is necessary to find some criterion for identifying an adequate hyperplane. One such criterion has been formulated for the method of reference vectors in the book: Vapnik V. N. and Chervonenkis A. Y. Pattern Recognition Theory. Moscow: Nauka, 1974. More precisely, many different criteria are considered in this book.
Both SVM and VMR are reference vector methods.
Which method is better or worse can be argued for a long time. However, you can take and check the generalizability and then everything will fall into place.
If the set is linearly separable, then the number of potential separating hyperplanes is infinite. In such a case it is necessary to find some criterion for identifying an adequate hyperplane. One such criterion has been formulated for the method of reference vectors in the book: Vapnik V. N. and Chervonenkis A. Y. Pattern Recognition Theory. Moscow: Nauka, 1974. More precisely, many different criteria are considered in this book.
Both SVM and VMR are reference vector methods.
Which method is better or worse can be argued for a long time. However, you can take and check the generalizability and then everything falls into place.
Problems should be solved as they arise, and putting the cart (model) ahead of the horse (predictors) is an absolutely futile exercise. All the more so to compare carts when it is not known what is harnessed to them and whether it is harnessed at all.
Before applying any type of models it is necessary to clear the list of predictors from noise, leaving only predictors that are "related" to the target variable. If you don't do this, you can easily slip into building models based on Saturn's rings, coffee grounds and other predictors that have been widely used in practice for several hundred years.
AboveDr.Trader tried to do the work to remove the noise from his set of predictors.
The result is negative.
I think that the reason of negative result is small number of observations at very big number of predictors. But this is the direction to dig before applying ANY models.
If the set is linearly separable, then the number of potential separating hyperplanes is infinite. In such a case it is necessary to find some criterion for identifying an adequate hyperplane. One such criterion has been formulated for the method of reference vectors in the book: Vapnik V. N. and Chervonenkis A. Y. Pattern Recognition Theory. Moscow: Nauka, 1974. More precisely, many different criteria are considered in this book.
Both SVM and VMR are reference vector methods.
Which method is better or worse can be argued for a long time. However, you can take and check the generalizability and then everything falls into place.
R has everything you need. See fTrading::sharpeRatio.
Oh, and PerformanceAnalitics wouldn't hurt to take a look.
Good luck