"New Neural" is an Open Source neural network engine project for the MetaTrader 5 platform. - page 36

 
TheXpert:

I don't know if the cognitron is anything like that.

Waiting for the sequel :)

The same as Fukishima's cognitron. HMAX and cognitron describe quite closely the functionality of different layers of the visual cortex.
 
TheXpert:

We need information on

-Conjugate gradient descent

-BFGS

In my time, I have studied in detail almost all methods of direct propagation network training. I'm sure that Levenberg-Marcadt method is the best(https://en.wikipedia.org/wiki/Levenberg%E2%80%93Marquardt_algorithm) out of gradient descent. It always finds a better minimum than all kinds of BACKPROPs and RPROPs, and even faster. What I posted on BPNN (some kind of RPROP) is child's play compared to LM. BFGS is more time consuming and the result is no better than LM's.
Levenberg–Marquardt algorithm - Wikipedia, the free encyclopedia
Levenberg–Marquardt algorithm - Wikipedia, the free encyclopedia
  • en.wikipedia.org
The LMA interpolates between the Gauss–Newton algorithm (GNA) and the method of gradient descent. The LMA is more robust than the GNA, which means that in many cases it finds a solution even if it starts very far off the final minimum. For well-behaved functions and reasonable starting parameters, the LMA tends to be a bit slower than the GNA...
 

I will now try to present my idea of building neural networks for analyzing price patterns. Those who have read my lectures 2 and 3 will immediately understand. The essence is the classification of price patterns by Buy, Sell, or Hold. The price at a certain period of time (say, 100 bars) is filtered by a layer of simple neurons S1. The input weights of these neurons describe the impulse characteristics of the filters. In the example of visual cortex, these weights described straight segments of different slope and length in two-dimensional image space. In quotes we also have a two-dimensional space: time and price. We can assume that S1 filter weights in our case also describe straight segments in time-price space of two possible inclinations: up and down. The slope angle depends on the length of each filter. These lengths can be pre-selected, e.g. 4, 8, 16, 32 bars. Each filter is a straight line normalized so that the sum of all values equals zero and the sum of squares equals 1 (or other normalization). In the next layer, let's call it S2, more complex patterns are formed from segments of layer S1, and so on. At the output of this multilayer transformation of quotes, we have a digital code describing the current pattern, the codes of patterns similar to each other, but differently stretched in time and price, are the same. These codes are fed to the inputs of the Support Vector Machine (SVM), which is pre-trained to determine Buy, Sell, or Hold conditions on historical patterns. The problem here is determining the shape of the filters in layers S1, S2, etc. I chose straight bars and their combinations for simplicity. By the way, in the model of visual cortex HMAX, all shapes of spatial filters are pre-selected based on biological experiments. We need to find an algorithm for automatically determining these filters. Such algorithms have already been developed for visual layer V1 (von der Malsburg, Linsker, Miller, LISSOM, Olshausen). We can borrow them for our task of classifying price patterns.

 
gpwr:
In my time I have studied in detail almost all methods of training direct propagation networks. I'm sure that of the gradient descents, the Levenberg-Marcadt method is the best(https://en.wikipedia.org/wiki/Levenberg%E2%80%93Marquardt_algorithm). It always finds a better minimum than all kinds of BACKPROPs and RPROPs, and even faster. What I posted on BPNN (some kind of RPROP) is child's play compared to LM. BFGS is more time consuming and the result is no better than LM.

I agree! In NeuroSolutions, for example, the Levenberg-Marcadt method converges where the other methods stand on local minima,

However, LM requires significant computational resources. The time per pass is longer

 
gpwr:

...

Each filter is a straight line normalized so that the sum of all values is zero and the sum of squares is 1 (or other normalization).

...

I have no proof now, but my intuition tells me that this double condition is contradictory,

the sum is 0 and the sum of squares is 1

will hold for a very narrow number of choices. If I'm wrong please kick me.

 
Urain:

I have no proof right now, but my intuition tells me that this double condition is contradictory,

the sum is 0 and the sum of squares is 1

will be executed for a very narrow number of choices. If I am wrong, please kick me.

No, it is not contradictory. For example, a 2-bar filter with values -0.707 and +0.707 satisfies the normalization conditions. The advantage of the first condition (sum of squares equals zero) is that it is not necessary to remove the mean from the quotes. The second condition (sum of squares equals zero) allows you to limit the sum of price1*filter1+price2*filter2 (that is our filter's output) within some range. The problem now is to define filter values (filter1,filter2,...,filterN) of arbitrary length N. It is possible to choose a straight line stretched between -filterMAX and +filterMAX, but this will be a simplification. With greater accuracy the filter will be similar to the principal component of quotations (PCA). But it will not be efficient. It would be more perspective if to search filter form as independent components of quotations (ICA) imposing condition of minimum number of such components (condition of sparsity). Later I will expose lecture 4 where I will explain everything.
 
TheXpert:

I see. But it is tied exactly to NN. It will be tied more accurately.

I'm afraid that it won't turn out something like "AWVREMGVTWNN" :) The main thing is to convey the essence, the nuances are not so important.

Meta Neuro Solution ?
 
Meta Neuro Engine (MNE)
 
gpwr:
Meta Neuro Engine (MNE)

Meta EngiNeuro (MEN) (c)

so we are engineers :)

 
Cool. I like it.
Reason: