• 沒有找到結果。

Proposed Niching SOMS Weight Updating Rule

Figure 4.3 shows the structure and operation of the SOM in the NSOMS. The SOM performs three operations: evaluation, deterministic competition, and search. In Figure 4.3, initially, we divide the whole SOM network into H subnetworks (niches), each niche comprises N neurons, and each neuron j in the hth niche (j ∈ Λh) contains a vector of a possible solution set whj (the weight vector). So the total number of neurons equals H × N in the whole network. The initial center location of the hth niche is set on the weh the average of all whj. Take the missile interception application as an example and let the number of incoming missiles be M. Each time new measured data [v1, v2, · · · , vM] are sent into the scheme, the SOM is triggered to operate. All of the possible solution sets in the neurons will then be sent to the dynamic model to derive their corresponding data phj. The SOM evaluates the product of all terms of °°°vm− phj°°° for m = 1, · · · , M . Of all the neurons for the hth niche, it chooses the neuron j, which corresponds to the smallest value, as the winner. The learning process then continues, and each niche will eventually converge to the nearest optimal solution.

The search strategy of the population-based optimization algorithm is to find the best individual and move other individuals to approach to the optimal solution. However, one drawback of the SOM-based optimization algorithms is that the network size increases exponentially along with the dimension (r) of the search space. The network needs least 2r neurons to ensure that each dimension can be considered during the search. To overcome this difficulty, an additional random term, such as random noise and random search methods, is added to raise the optimization efficiency. It might then use few neurons and randomly explore to each coordinate direction during the search. In [40], a small amount of random noise and also a narrowing down method are included in the weight updating rules

to improve its performance. In [29], Michele et al. also derive an alternative optimization algorithm based on neural gas networks (NG-ES) to overcome the bad scaling problem of the KSOM-ES by introducing a mechanism for generating trial points randomly. Wu and Chow proposed a self-organizing and self-evolving agents (SOSENs) neural network that combines multiple simulated annealing algorithms (SAs) and SOM algorithm [44]. Each neuron of SOSENs has its own updating rules (self-evolving) with an SA, and learns from other neurons by the SOM algorithm (self-organizing) after some time. However, when the distance between the best current solution and the real optimal solution is very large, the search process of these methods do not achieve good performance with only small random changes.

From the discussions above, the SOMS weight updating rule previously proposed may not be suitable for optimization in a multimodal domain. Thus, we made several modi-fications so as to reduce the number of neurons to raise the optimization efficiency. We similarly define a Gaussian distribution function G(whj,i(k)) as distribution function for each element whj,i(k), the ith element in whj(k) in the kth stage of learning:

G(whj,i(k)) = exp(−(wj,ih (k) − whei(k))2

hi(k)2 ) (4.1)

where whei(k) stands for the ith element in whe(k) average of all whj(k), and σhi(k) is the standard deviation of the distribution for whj,i(k). From the same concept to speed up the learning, described in chapter 3, the strategy is to vary the mean and standard deviation of G(whj,i(k)) by moving its center toward whj,i(k) and enlarging (reducing) the standard deviation σih(k) according to the double distance between whj,i(k) and whei(k). The new distribution function ˜G(j,ih (k)) is then formulated as

G( ˜˜ whj,i(k)) = exp(−( ˜wj,ih (k) − ˜whei(k))2

σhi(k)2 ) (4.2)

where ˜whj,i(k) stands for the new whj,i(k), ˜whei(k) the new whei(k), and ˜σih(k) the new σih(k) after the adjustment. Based on the same strategy in the SOMS, during each iteration of learning, G(whj,i(k)) is dynamically centered at the location of the winning neuron j, with a larger (smaller) width when wehi(k) is much (less) different from whj,i(k). It thus covers a more fitting neighborhood region, and leads to a higher learning efficiency.

In order to reduce the network size greatly for dealing with the optimization of a multimodal domain, we make the adjustment in weight updating rule of the SOMS.

From the mapping property of the SOM, we understand that the SOM cannot obtain a good feature maps with a small network size. Therefore, let the weight vectors form the uniform distribution like the pre-ordered lattice in the neuron space becomes not so meaningful. So far a more fitting search range is already well-defined with the determined mean and standard deviation of the Gaussian distribution function. Thus, we propose a deterministic neighborhood to design the NSOMS weight updating rules. Based on the proposed concept, the new ˜whei(k) and ˜σhi(k) are then formulated as

˜

wehi(k) = wehi(k) + ηw(k) · (wjh,i(k) − whei(k)) (4.3)

˜

σih(k) = σhi(k) + ηw(k) · (2|whj,i(k) − whei(k)| − σih(k)) + ε (4.4)

where ηw(k), (0 < ηw(k) ≤ 1) stands for the learning rate in the kth stage of learning and ε the a small value added to avoid that σih(k) rapidly converges to zero. We can set a large value of ηw(k) to speed up convergence. However, if ηw(k) set to be 1, it

may probably converges to the local optimum. Thus, the premature convergence can be avoided through introduction of the additional adaptation term. From Eqs.(4.3) and (4.4), we can regenerate the new weight ˜whj,i(k) from a Gaussian distribution with mean

˜

Under this learning process,the network will gradually converge to a very small region with σhi(k) continuing to decrease.

Sometimes more than two niches eventually converge to the same location of the optimal solution, or several optimal solutions have not been found yet. To overcome this difficulty, a technique for automatically determining the number of niches is introduced into the NSOMS to find as many solutions as possible. First, we use Eqs.(4.3) and (4.4) to detect a searched optimal solution when the standard deviation ˜σhi(k) for every element is less than preset value and to determine an effective optimal solution with duplicate optimal solutions excluded when the mean values ˜wie(k) and ˜wje(k) are very close. If more than two niches are similar, only one is reserved and the others eliminated from the competition. If the number of niches is equal to the number of effective optimal solutions, a new niche is generated randomly. In other words, we intend to make the niche set size H(k) vary depending on the effective optimal solutions set size Es(k). We thus define a specific relation between the niche set size H(k) and the effective optimal solutions set size Es(k).

H(k) = n1· Es(k) + n0 (4.6)

where n0 and n1 are positive integers which can be either constants or variables decaying along with time. Of course, other types of functions can also be used. During the searching process, in order to prevent that the new niches regenerated converge on the locations of searched optimal solutions repeatedly, the initial center location of the regenerate niche should be far away from the initial center locations of all previous niches as much as possible. Hence, we define an evaluation criterion as

DWi∗ = min

i

°°

°Rh− wCi°°°≥ λ(k) (4.7)

where ˜wRh stands for the initial center location of the regenerated niche, wCi the ith location included in the set <C of the initial center locations of all previous niches, and λ(k) the distance evaluation parameter. If the minimum distance DWi∗ less then λ(k), this new niche will be regenerated randomly again. In the initial stage of the learning, λ(k) can be set larger to prevent that the similar niches converge to the same optimal solution repeatedly. Later, λ(k) may be decreased gradually to let some optimal solutions that are possibly very close can found. Thus, a function for λ(k) that satisfies the demand can be formulated as

λ(k) = λ(0)

2 · e−k/τ (4.8)

where the initial value λ(0) is the average of all kwi(0) − we(0)k and τ time constant. Of course, other types of functions can also be used. Under this design, the searching efficacy

of the NSOMS will become faster and more efficiently.

4.3 Visualization Of the Distribution Of Optimal So-lutions

Different optimal solutions can now be found by the NSOMS. The next significant task is how to select the most useful solutions from the set of the optimal solutions. Visualization and clustering of high-dimensional data are well-known successes in SOM. We employ the basic principle of the double self-organizing map (DSOM), which updates the weight vectors together with the two-dimensional position vector of the neuron, to achieve the visualization of distribution of optimal solutions. In other words, the positions of the optimal solutions within the parameter space are mapped onto a two-dimensional (2-D) space. Through this map, it allows us to classify the optimal solutions into clusters easily, yielding useful information for solution selection.

The NSOMS is different from the DSOM in that the position updating rules of the neurons in the DSOM cannot be applied directly to optimization due to the system parameters operate in quite different ranges. We thus proposed a new adaptive mapping model to visualize the distribution of the optimal solutions. First, we know that the neuron space and weight vector spaces are with different dimensions, so we have to transform them into the same dimension. We define two Gaussian type functions as the neighborhood functions in the neuron space and the weight vector space, Di and F (wi(k)) in the kth stage of learning as

Di = exp(− niche, respectively, wi(k) the weight vector of neuron i to entire network, σd the average of all°°°ri− rhj

°°

°2, and σcthe average of all°°°wi(k) − whj(k)°°°2. The Gaussian type function is frequently used as the neighborhood function, and it is differentiable and continuous.

With the neighborhood functions, the magnitudes of their distances in the neuron space and weight vector space, respectively, can be normalized to be between 0 and 1. The new learning model in the SOM is designed to let nearby neurons of a feature map correspond to nearby weights. From this mapping, a cost function Ei(k) is then defined as

Ei(k) = 1

2(Di− F (wi(k)))2. (4.11) Based on the gradient-descent approach, the position updating rule of the neurons is derived as

ri(k + 1) = ri(k) − ηp(k)∂Ei(k)

∂ri(k)

= r. i(k) + ηp(k)(Di· (Di− F (wi(k))) · (ri(k) − rhj(k))) (4.12)

where ηp(k) stands for the learning rate in the kth stage of learning.

It is evident that only two learning parameters ηw(k) and ηp(k) need to be deter-mined. It is then straightforward to determine the learning parameters for the diverse optimization problems. Through the weight and position updating rules of the neurons

synchronously, the NSOMS can be applied to optimization, in particular, identification and visualization of multiple optimal solutions on the 2D neuron space.

4.4 Applications

To demonstrate its capability, the NSOMS is applied to both function optimization and dynamic trajectory prediction. A PC with 3GHz and MATLAB software were used for all the simulations. Based on the NSOMS, we first develop learning schemes corresponding to each of the applications. Simulations are then executed for performance evaluation.

The results are especially compared with the RCS-PSM [18] for their similar searching abilities. To compare their performances, the maximum ratio of the searched peak values to the real peak values (MPR), which denotes the sum of the fitness of the searched optima divided by that of the fitness of the actual optima and the effective number of maintained peaks (NMP), are selected as the performance criteria [18]. A searched peak is detected with the best fitness value at least 80% of the real peak value.

4.4.1 Function Optimization of a Multimodal Domain

For a function optimization problem, the goal may be to maximize (minimize) an object function O(·). Let O(whj(k)) be the function value for the weight vector whj(k), which represents a possible solution. During the learning process, Pr is a reference value of the performance evaluation for optimization. Note that the value of Pr is empirical according to the demanded resolution in learning, and we chose it very close to zero. The learning algorithm for function optimization is organized as follows.

Algorithm for function optimization based on the NSOMS: Maximize (minimize) an object function using the NSOMS.

Step 1: Set the stage of learning k = 0. Choose H number of niches, N number of neurons within each niche, and reference value Pr. Estimate the ranges of the possible parameter space and randomly store the possible parameters whj(0) into the neurons, where j = 1, . . . , N , h = 1, . . . , H.

Step 2: Compute O(whj(k)) for all whj(k).

Step 3: Among the neurons to every niche, find the one with the largest (smallest) value as the winning neuron j for the maximization (minimization) problem.

Step 4: Update the weight vectors of the winning neuron j and its neighbors to every niche, and update the positions of neurons to entire network according to the position and weight updating rules described in Eqs.(4.3)-(4.5) and Eq.(4.12).

Step 5: If Pq

i=1σ˜hi(k) < Pr to every niche, whj(k) is determined to be as an effective optimal solution with duplicate optimal solutions excluded. From Eqs.(4.6) and (4.7), the new whj(k) will be randomly regenerated and then added to the set <c.

Step 6: Check whether the number of iterations is smaller than a pre-specified maximum number of iterations. If it is not, let k = k + 1, and go to Step 2; otherwise, the learning process is completed and output all optimal values. The final network mapping is ready for visualizing the distribution and structure of the optimal solutions.

Four multimodal functions are used to demonstrate the proposed algorithm. These functions are defined as follows:

F 1(x) = sin6(5πx), where 0 ≤ x ≤ 1 (4.13)

These four test functions have also been used in [15, 18]. The optimization here is to maximize these four functions. The first function, F1, has five peaks with the same height for x = 0.1, 0.3, 0.5, 0.7, and 0.9. The F2 function has five unequally spaced peaks with different heights. In F3, a(i) = 16[(i mod 5)−2] and b(i) = 16([i/5]−2). F3 is a two-dimensional function with 25 peaks of different heights in the interval [476.191, 499.002]

and the highest peak is located at (32,32). The 2D test functions F1-F3 are shown in Figure 4.4. F4 is a four-dimensional function with 4 peaks of the same height, where yi is a constant vector obtained by permuting the components of the vector (8,2,2,2).

The initial number of the neurons was set to be 1 × 10 (H = 1, N = 10) and the pa-rameters n1, n0, ηw(k) and ηp(k) were set to be constants, 2, 1, 0.4, and 0.4, respectively for all function optimization. For comparison, we also use the RCS-PSM for function minimization, which is with an initial population size of 10 to match with that of the SOM, and the crossover and mutation probability of 0.6 and 0.0333 for the GA opera-tion, respectively. To automatically determining the number of niches, we use the same condition described in Eq.(4.6) to RCS-PSM for simulations.

For each test functions, each algorithm is repeated 30 times and carried out after 30 iterations each time. The comparison results of average are listed in Table 4.1. As shown in Table 4.1, NSOMS and RCS-PSM perform quite well in terms of the number and quality

0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

Function F3(x,y) Function F2(x)Function F1(x)

of peaks obtained. But, NSOMS works better than RCS-PSM for the computational time.

Figure 4.5 shows the variations of NMP and MPR according to the increase of time for F3 test function. We observed that the NSOMS converged faster than the RCS-PSM.

Figure 4.6 shows the variation of the network structure on the SOM using the NSOMS for F4 test function. In Figure 4.6(a), we only show the variation of the best four niches.

It shows clearly that the positions of neurons on the SOM reveal the distribution and structure of the optimal solutions. Figure 4.6(b) shows the neighboring relationship of the neurons of the best four niches, and we can observe that the neighborhood functions of neurons and weights, described in Eqs.(4.9) and (4.10), varied during the NSOMS learning process, and eventually they are very close to each other. From these results, the NSOMS demonstrates the identification and visualization of the optimal solutions of high dimension.

Table 4.1: Comparison results for NSOMS and RCS-PSM on the 4 test functions.

Function Method NMP MPR Time (s)

NSOMS 5 1 0.1213

F1

RCS-PSM 5 1 0.2154

NSOMS 5 1 0.1436

F2

RCS-PSM 5 1 0.2417

NSOMS 25 1 2.0301

F3

RCS-PSM 25 1 5.5171

NSOMS 4 1 0.1337

F4

RCS-PSM 4 1 0.7265

(a) Number of maintained peaks

(b) M aximum peak ratio

Figure 4.5 Convergence comparisons for F3 function: (a) the variation of the number of maintained peaks and (b) the variation of the maximum peak ratio during the NSOM S learning process.

(a) Projection result in 2D neuron space

(b) Final neighborhood function values

Figure 4.6 The results obtained by the NSOM S for F4 function: (a) projection result in the 2D neuron space, and (b) final neighborhood function values.

:Di :F(wi( ))k

:Di*

4.4.2 Multiple Dynamic Trajectories Prediction

For a dynamic trajectory prediction problem, the goal may be to estimate the initial position and velocity of a moving object using the measured data. For the trajectory prediction of multiple targets, here we assume that the target detection has been carried out in advance, and we focus on the estimation of the initial states of multiple moving objects. Through a learning process, the NSOMS may determine a most probable initial state of each target through repeatedly comparing the measured data with the predicted trajectories derived from the possible initial states stored in the neurons of the SOM.

In this application, the nonlinear dynamic equation describing the trajectory of the moving object and the measurement equation are as those described in Chapter 3. The learning algorithm for multiple trajectories prediction is organized as follows.

Algorithm for multiple trajectory prediction based on the SOMS: Predict an optimal initial state for the trajectory of every moving target using the measured position data.

Step 1: Set the stage of learning k = 0. Choose H number of niches, N number of neurons within each niche, and reference value Pr. Estimate the ranges of the possible position and velocity of the moving object, and randomly store the possible initial states whj(0) into the neurons, where j = 1, . . . , N , h = 1, . . . , H.

Step 2: Send whj(k) into the dynamic model, described in Eqs.(3.25) and (3.28), to compute phj(k).

Step 3: For each neuron j of every niche, compute its output Ojh(k):

Ohj(k) =YM where M is the number of the objects detected.

Find the winning neuron j with the minimum Ohj(k):

Ohj(k) = min

j Ohj(k) (4.18)

Step 4: Update the weight vectors of the winning neuron j and its neighbors to every niche, and update the positions of neurons of the entire network.

Step 5: If Pq

i=1σ˜hi(k) < Pr for every niche, whj(k) is determined to be as an effective optimal solution with the duplicate optimal solutions excluded. The prediction process outputs the predicted optimal initial states to the dynamic model to derive the object trajectories. From Eqs.(4.6) and (4.7), the new whj(k) will be randomly regenerated and then added into the set <c.

Step 6: Check whether the number of iterations is smaller than a pre-specified maximum number of iterations. If it is not, let k = k + 1, and go to Step 2; otherwise, the prediction process is completed and output optimal states of all objects. The final network mapping provides the visualization of the distribution of the optimal states. In addition, during each stage of learning, we perform a number of learning to increase the SOM learning speed. This number of learning is set to be a large number in the initial stage of the learning process, such that the NSOMS may converge faster at the price of more oscillations, and decreased gradually to achieve smooth learning in later stages of learning.

To demonstrate the effectiveness of the proposed NSOMS and weight updating rule, we performed a series of simulations for dynamic trajectory prediction based on using the proposed NSOMS and two NSOMS without the proposed dynamic weight updating rule (named as SOMSO-1 and SOMSO-2 by using the SOMO’s and SOSENs’s weight updating rules, respectively, in place of that of the NSOMS). The trajectory to predict in the simulations was designed to emulate that of a missile. Its governing equations of motion in the 3D Cartesian coordinate system are as those described in Chapter 3. The ranges of the possible initial states wj(0) were estimated to be

1.14 × 106m ≤ x1(0) ≤ 2.14 × 106m

Within the ranges described in Eq.(4.19), the possible initial positions and velocities of the missile were selected and stored into the 1125 (5×225) neurons of the 2D SOM. We consider three targets to be detected in the following simulations. The parameters of NSOMS were set to be n1 = 2, n0 = 1, respectively. The additional adaptation term ε, described in Eq.(4.4), was set to be 0.1. For comparison, we set the same learning rate as those in the NSOMS described in Sect. 4.4,1, and several parameters of the SOMSO-1 and SOMSO-2 were adjusted via a trial-and-error process to yield salient performance.

The number of learning is set to be 20 during each stage of learning.

We first applied the SOMS, SOMSO-1, and SOMSO-2 for trajectory prediction with a good estimate of the initial state. Three ideal initial states of the missiles were assumed

We first applied the SOMS, SOMSO-1, and SOMSO-2 for trajectory prediction with a good estimate of the initial state. Three ideal initial states of the missiles were assumed

相關文件