Develop a methodology for phase-space prediction

[Fig. 1b]. A phase-space plot of the Henon map and a white noise sequence in Fig. 2, on the other hand, reveals remarkable structure in the chaotic Henon map while the white noise sequence fills up the entire plane with no apparent structure.

2.2 Develop a methodology for phase-space prediction

Once we have reconstructed the phase space, we can use some of its properties to develop a short-term prediction model. For example, if the underlying dynamics is determi- nistic, then the order with which the points in the phase space appear will also be deterministic. Thus, we may be able to define some functional relationship between the current state Xt and the future states Xt þ P, i.e. Xt þ P ¼ F p Xt. Now, we need to find a predictor F p that approximates F p . There are a variety of numerical tech- niques to approximate F p from scattered points in the phase space. This methodology can be illustrated by using Fig. 3, where part of a trajectory is shown in a two- dimensional phase-space and the present state is denoted by an open circle. The solid circles indicate neighbors of the current state, and the arrowheads show movement of the neighbors through a local section of the phase space. By finding a suitable function linear or nonlinear that approxi- mates how the neighbors move, a prediction of the current state can be made. This is know as local approximation as opposed to a global approximation which defines a functional relationship over the entire phase space. Farmer and Sidorowich 6 introduced local linear models for phase-space forecasting. Smith 2 discussed the relation- ship between local linear and nonlinear models as well as between the local and global approaches. In general, local linear approximation has been shown to provide better prediction accuracy for a number of controlled datasets 18 . In this paper, local approximation methods will be used. One such method, popularly known as the nearest neighbor method, approximates unknown functions near the present Fig. 1. a Time series of 100 points for the chaotic Henon map: x t þ 1 ¼ 1 ¹ ax 2 t þ y t ; y tþ 1 ¼ bx t with a ¼ 1.4 and b ¼ 0.30. This time series is in many ways indistinguishable from random noise. b Time series of 100 points generated from uniform distribution in the interval between 0 and 1. Fig. 2. Two-dimensional phase space map for a a Henon map, and b a white noise sequence. Phase-space analysis of daily streamflow 465 state vector by using the nearest neighbor of the present state. We now locate nearby M-dimensional points in the phase space and choose a minimal neighborhood with K closest neighbors such that the predictee the point from which the prediction is made is contained within the smallest simplex. To enclose a point in an M-dimensional space, we require a simplex with a minimum of M þ 1 points. Then, to obtain a prediction, we project the domain of the chosen nearest neighbors T P prediction step steps forward and compute F p to get the predicted value. Since it becomes increasingly difficult to define an enclosing simplex for higher dimen- sional embedding spaces, we have extended the above idea to the nearest neighbors in an Euclidean sense. A minimum of M þ 1 nearest neighbors are chosen based on the Euclidean distance between the neighbor and the predictee. Then, we project the domain of the chosen neighbors T p step forward and estimate the predicted value. We have explored several estimation kernels including arithmetic average, weighted average and weighted regression to estimate the predicted value. It was found that arithmetic average provides comparable prediction accuracy and requires no tuning parameters and hence we have chosen arithmetic average of projected neighboring points to obtain the predicted value in this study. There are only two parameters to be chosen for this phase-space prediction model: embedding dimension M, and number of nearest neighbors K. In general, M min . 2D þ 1 where D is the attractor dimension. An estimate of the attractor dimension may be obtained from the corre- lation dimension 4,5 . Prediction results are sensitive to the choice of M 10 . We will look at the prediction accuracy correlation between predicted and observed as a function of embedding dimension to choose an optimum value of M for our prediction algorithm. Since to enclose a point in an M -dimensional space, we require to construct a simplex with a minimum of M þ 1 points, one has K min . M þ 1. Use of the phase space to develop a forecasting model may appear to be similar to an autoregressive model: a pre- diction is estimated based on time-lagged vectors. However, the crucial difference is that understanding phase-space geometry frames forecasting as recognizing and then repre- senting underlying dynamical structures. For example, two neighboring points in a phase space may not be close to each other within the context of a time sequence. The traditional autoregressive AR model relies on time-lagged signals that are neighbors in a temporal sense, whereas a neighbor in a phase space is close in a dynamic sense. In addition, once the number of lags exceeds the minimum embedding dimension, the geometry of the underlying dynamics will not change. A global linear model, such as the AR model, must do this with a single hyperplane with no fundamental insight into the underlying geometric structure. Unlike traditional AR models, the proposed methodology also promises to make a tentative distinction between stochastic noise and low-dimensional chaos. A characteristic feature of chaotic dynamics is that the prediction accuracy exponen- tially decays as the prediction time increases. On the other hand, for a noisy system the prediction accuracy does not decay sharply with prediction lead time 9,10 .

2.3 Distinction between deterministic chaos and stochastic noise