Machine learning in trading: theory, models, practice and algo-trading - page 596

You are missing trading opportunities:
- Free trading apps
- Over 8,000 signals for copying
- Economic news for exploring financial markets
Registration
Log in
You agree to website policy and terms of use
If you do not have an account, please register
Try cross-validation (K-fold).
How does it help strengthen the impact of fresh data?
Think about it, you teach the model by feeding separate blocks of data, which will give the model some independence from time series sequence and lead to the fact that the new data will be evaluated without "bias".
"independence from the sequence of the timeseries" is provided by mixing. Without it, the model comes to nothing at all.
And the question is about how to increase the importance of the freshest data when mixing, so that the model picks up new market trends faster.
"independence from the sequence of the timeseries" is provided by mixing. Without it, the model comes to nothing at all.
And the question is how with mixing to increase the importance of the freshest data, so that the model would grasp new market trends faster.
I.e. training in 2 steps?
Training on a large amount of data + retraining of the obtained model on fresh data.
It is possible to try.
I had an idea, just add fresh data 2 -3 times to the total training data. Even with shuffling, the significance will increase.
So I thought, if everything is shuffled, how can we make the fresh data have a stronger effect on the training?
There is a trick to duplicate the most recent training examples several times.
And for example in gbm package you can set some coefficient of importance of each training example, but it's not a neuron, I just gave an example.
"Sequence-independent timeseries" is provided by shuffling. Without it, the model comes to nothing at all.
In most models, there is no such thing as sequence dependence at all. In neurons, for example, an error is calculated for each training example, and then the sum of all errors affects changes in weights. The sum does not change due to the change of places of the summands.
But models often have the batch.size parameter or similar, it affects what percentage of training data to take for training. If you take a very small percentage of training data, and turn off shuffling, then the model will take the same small set every time, and everything will end badly. I don't know about darch specifically, but turning off mixing shouldn't cause a complete failure, you have something wrong with the other parameters.
Try cross validation (K-fold).
I fully support that. Whatever loud claims the author of the model makes about its protection from overfit, only k-fold will show if it's true.
It's time for you to round up and make a conclusion.
Coming soon... "almost done."
I have never done such a tin before in my life
coming soon... "almost done."
I've never done such a tin before in my life.