In state feedback pole placement, the state feedback gain must be determined for a given system such that the closed-loop poles coincide with the desired locations. This is a well-known problem, and the pole placement methods have been extensively discussed in the literature     . In standard pole placement methods, a state space model is assumed to be given by a system identification technique using data from past experiments. Whereas the traditional approach combines the identification of the state space model with the standard pole placement method; an alternative approach called “data-driven pole placement” has recently been proposed  . In this approach, the state space model and pole placement feedback gain are identified simultaneously from the set of state measurements and control input sequences. The method proposed in  is based on the data-driven control framework (  and references therein) such as unfalsified control  , virtual reference feedback tuning (VRFT)   , or fictitious reference iterative tuning (FRIT)     . In the data-driven control framework, where no explicit mathematical plant model is used, a feedback controller must be derived that satisfies the prescribed closed-loop performance and fits to known experimental data. In contrast with traditional model-based controller designs, techniques such as controller identification  or a combination of plant model and controller identification must be applied   .
Many studies of data-driven control have focused on output feedback control and data-driven state feedback control    , in which the prescribed closed-loop performance is achieved by applying a closed-loop reference transfer function. Such methods can be applied to the data-driven pole placement problem by choosing a reference transfer function with the desired poles. However, the zeros of the reference transfer function cannot normally be specified, because the zeros of the plant are unknown. In contrast, the data-driven pole placement method presented in  requires only a state space representation of the closed- loop system to specify the prescribed closed-loop performance, as shown in Section 2. This avoids the zero assignment issue that arises in the transfer function approach used in  .
This data-driven pole placement method can, therefore, be applied to linear and time-invariant systems with measurable states. The method is briefly reviewed in Section 2. However, the capacity of the data-driven pole placement method to handle noise remains an open issue, though in  , the total least square (TLS) method  was claimed to be eﬀective. Measurement noise is one of the issues which may surely face in practical applications. Therefore, to resolve this, we introduced a prefiltering technique that reduces the eﬀect of measurement noise in Section 3. More specifically, a finite impulse response (FIR) filter is used to prefilter the data, as this makes them easier to manipulate. In Section 4, by using the numerical example of a self-balancing robot, we discuss the eﬀect of applying this prefiltering technique, together with the least square (LS) and TLS methods, to a self-balancing robot model. We investigate the ability of the data-driven pole placement method to produce a linearized model using numerical simulations as in  . A nonlinear diﬀerential equation was used to represent the dynamics of a self-balancing robot there. Moreover, we evaluate the effects by two different exciting signals, the random and the chirp exciting signal, along with TLS and prefiltering. Finally, we compare all the results for the pole placement error and identification error when two exciting signals are applied.
Notation: Let and be and matrices, respectively. Then, the Kronecker product of and is a matrix, defined as follow:
where is the element of . The vectorization of then stacks the columns into a vector:
in which is the column of . The Frobenius norm of matrix is defined as
2. Data-Driven Pole Placement
In this section, we briefly review the data-driven pole placement method formulated in  .
Consider a discrete-time linear time-invariant system and static state feedback
where is the state vector, is the input vector, is the feedback gain, and is the external input to the closed loop system.
The data-driven pole placement problem was formulated in  as follows:
Problem 1. We assume that the order of the plant n is known, state n is measurable, pair is controllable but the exact value is unknown and is of full rank. Let be a self-conjugate set of n complex numbers in the unit circle. Given the input and output measurement data sequence of (4), find a state feedback gain from the observed data such that
In a conventional approach, this problem is solved in two steps: and are identified from , then is derived using the standard pole placement algorithms. In contrast, the data-driven pole placement method solves the two steps simultaneously. To achieve this, the method uses the equivalency between the closed-loop system
with the desired pole placement gain and
where with is an appropriate controllable pair. This equivalency requires the nonsingular matrix to exist. Then, we remove from (7) by using (5), to obtain
Then, using (8), we obtain
If satisfies (10),
In  , Equation (11) is cast into
Remark 1. The system in (7) can be interpreted as a reference model within VRFT (e.g.,   ) and FRIT (e.g.,     ). The idea of eliminating in (9) is also based on FRIT. In    , a similar state feedback control problem has been discussed within the FRIT framework. To apply these FRIT techniques to the data-driven pole placement problem, the desired transfer function must be specified from to , rather than . When precise values for are not available, it becomes impossible to specify the zeros of the desired transfer function.
Remark 2. To obtain the datasets in (12) by applying state feedback in (5) to the system in (4), the initial feedback gain should be based on . Hence, in Problem 1, the exact value of is assumed to be unknown.
When applying the property of Kronecker product (see for example Th.2.13 in  ) to the transpose of (15) to solve (15) for and , a further linear equation is derived, as follows:
If is nonsingular, the model coefficients can be obtained
3. Prefiltering Noisy Measurement
When the measurement of is contaminated by noise ,
Then, (10) becomes
Hence, if satisfies the above equation,
Then, the resulting linear equation is given as
where the effect of noise has the same structure as in (20), then
and is the equation error. Following  , we can solve to (27) as a TLS problem  , by minimizing the Frobenius norm . It is known that the TLS solution is given as
based on the singular value decomposition
where these matrices are partitioned into blocks corresponding to and .
Here, we assume that there exists such that
for all . This means that when
for the matrix
where each column has elements of 1. Therefore,
This multiplication by represents the prefiltering of signals via an order FIR filter.
When the systems (4) and (7) are driven by the exciting signal, we have
By applying to these systems, we obtain
Here, if , (34) cannot be satisfied. Hence, for all i, , that is
must be satisfied.
4. Numerical Example: Self-Balancing Robot
We next applied the data-driven pole placement method described above to the model of a self-balancing robot   as shown in Figure 1. The robot is equipped with right and left wheels driven by direct current (DC) motors whose voltages and can be controlled. Because the motion dynamics can be decomposed by the input , the control input to the robot was represented as
Figure 1. (a) Coordinates of the self-balancing robot; (b) photo.
We assume that the pitch angle and the pitch angular velocity of the body could be measured, as well as the angles and of the right and left wheels, and their angular velocities and , respectively. We define the mean values of the right and left wheel angles and , and the yaw angle of the body as follows:
where is the radius of the wheel and is the distance between the two wheels.
4.1. Equation of Motion
The equation of motion for the self-balancing robot can be derived as
The symbols are explained in Table 1. The parameters used in the simulations were taken from   .
Table 1. Parameters of the self-balancing robot   .
4.2. Linear Model and Feedback Gain
We linearized the equations of motion (50) around equilibrium states , , , , , , and . Then, under the assumption that , , , , , and , the linearized equations of motion can be derived as
By defining the state vector
the linear state space model can be derived as
Then, the feedback can be independently designed as
Note that this can be more succinctly represented as
When the parameters in Table 1 are used and the sampling period is , the discrete-time model after discretizing (55) is
Here, we assume that the exact values of (60) are not available, but that uncertain values are available:
The coefficients can be derived from , with an assumed uncertainty of 10%. By applying linear quadratic optimal control theory to (61), the desired closed-loop pole locations can be chosen as
and the initial feedback gains needed to obtain datasets for the data-driven pole placement as
4.3. Comparison of Methods
Next, simulations were conducted and comparisons were made from the obtained results when using different methods and exciting signals.
Measurement noise was prepared with the Gaussian distribution , where , and in , , and , respectively. This is shown in Figure 2(a). We used the random exciting signal shown in Figure 3(a) and the linear chirp signal shown in Figure 3(b) with the uniform distribution and . We set the order of the prefilter (33) as . After prefiltering, the measurement noise in , , and was reduced, as shown in Figure 2(b). The prefiltered exciting signals were shown in Figure 3(b) and Figure 3(d). It can be seen that the exciting signals were not eliminated by prefilter , but that the high-frequency elements were reduced.
A closed-loop response in the presence of measurement noise by state feedback (56), with initial gain (64), is shown in Figure 4. The response to the random exciting signal and the chirp exciting signal are shown in Figure 4(a) and Figure 4(b), respectively. Of particular note is that the responses of , , and in Figure 4(b) show the high-pass filter-like gain characteristics of the transfer function from to .
For comparison, the dataset for the data-driven pole placement was chosen as where and .
To evaluate the obtained pole placement gain , we introduced an accuracy measurement that takes the largest absolute difference in value between each eigenvalue of and the corresponding ,
Figure 2. (a) Measurement noise; (b) Prefiltered measurement noise.
Figure 3. Exciting signal (a) random, (b) chirp, (c) prefiltered random, (d) prefiltered chirp.
Figure 4. Closed-loop response by an initial state feedback via (a) random exciting signal (b) chirp exciting signal .
To evaluate the obtained model , the following identification errors were used:
The eigenvalues were sorted by magnitude using the MATLAB command “sort”. This further sorts elements of equal magnitude by the phase angle on the interval . The impulse response was used to evaluate the model obtained, as follows:
where and are the impulse responses of and , respectively.
From the perspective of system control, smaller is better, particularly in the case of , , and . The following key results were contrastively found in Table 2:
1) The initial model and feedback gain were affected by uncertainty: The model errors and pole placement errors are shown in Table 2 (initial).
2) The results when using the LS method to solve linear Equation (27) for noiseless data are shown in Table 2(a). All errors were reasonably small, confirming that the data-driven method performs well when the measurement data are noiseless.
3) The results when using the LS method to solve linear Equation (27) for noisy data are shown in Table 2(b). All errors became larger when noise was added, suggesting that LS analysis is inadequate when the measurement data are contaminated by noise.
4) The results when using the TLS method to solve linear Equation (27) are shown in Table 2(c). The errors were significantly smaller than those reported in  , using the LS method.
5) The results when applying prefiltering (PF) and using the TLS method to
Table 2. Comparison of errors.
Figure 5. Comparison of pole locations (“+” indicates the desired poles, “.” those obtained by the random exciting signal and “o” those obtained by the chirp exciting signal.).
solve linear Equation (27) are shown in Table 2(d). The prefilter further reduced the errors, in particular, the pole placement error and the impulse response error .
6) The results when applying PF and using the TLS method to solve the linear Equation (27), but with as the chirp signal, are shown in Table 2(e). No significant improvement in error rates was found with respect to when using the chirp exciting signal. However, the errors with respect to became significantly worse than when a random exciting signal was used. This was assumed to be because has an unstable eigenvalue of 1.7838. We conclude that a random exciting signal is more appropriate than a chirp exciting signal when using data-driven methods.
Finally, we compare the pole locations obtained as shown in Figure 5. As can be seen, a better performance was achieved when using the random exciting signal.
In this study, we evaluated the different approaches reducing the effect of measurement noise in data-driven pole placement methods for deriving a state space model and pole placement state feedback. Using numerical simulations of a self-balancing robot, which is a nonlinear system, we demonstrated the important role that prefiltering can play in reducing the interference caused by noise. Again using numerical simulation, we compared the use of two exciting signals: a random signal and a chirp signal. The use of a random exciting signal was found to be more effective with our proposed method. Further developments are needed in the methods used to cope with noise. A method such as that used in  may be appropriate for use in practical applications where noise is present, and adaptive control based on real-time updating  is a future promising approach.
This work was partially supported by JSPS KAKENHI Grant Number 16H04385.
 Hikita, H., Koyama, S. and Miura, R. (1975) The Redundancy of Feedback Gain Matrix and the Derivation of Low Feedback Gain Matrix in Pole Assignment. The Society of Instrument and Control Engineers, 11, 556-560. (In Japanese)
 Yamamoto, S., Okano, Y. and Kaneko, O. (2016) A Data-driven Pole Placement Method Simultaneously Identifying a State Space Model. Transaction of the Institute of Systems, Control and Information Engineers, 29, 275-284. (In Japanese)
 Souma, S., Kaneko, O. and Fujii, T. (2004) A New Method of a Controller Parameter Tuning Based on Input-output Data-Fictitious Reference Iterative Tuning. Proceedings of the 2nd IFAC Workshop on Adaptation and Learning in Control and Signal Processing, 37, 789-794.
 Matsui, Y., Akamatsu, S., Kimura, T., Nakano, K. and Sakurama, K. (2011) Fictitious Reference Iterative Tuning for State Feedback Control of Inverted Pendulum with Inertia Rotor. SICE Annual Conference, Tokyo, 13-18 September 2011, 1087-1092.
 Matsui, Y., Akamatsu, S., Kimura, T., Nakano, K. and Sakurama, K. (2014) An Application of Fictitious Reference Iterative Tuning to State Feedback, Electronics and Communications in Japan, 97, 1-11.
 Kaneko, O. (2015) The Canonical Controller Approach to Data-Driven Update of State Feedback Gain. Proceedings of the 10th Asian Control Conference 2015 (ASCC 2015), Kota Kinabalu, 31 May-3 June 2015, 2980-2985.
 Van Heusden, K., Karimi, A. and Soderstrom, T. (2011) Extensions to “On Identification Methods for Direct Data-Driven Controller Tuning”. International Journal of Adaptive Control and Signal Processing, 25, 448-465.
 Kaneko, O., Miyachi, M. and Fujii, T. (2008) Simultaneous Updating of a Model and a Controller Based on the Data-Driven Fictitious Controller. 47th IEEE Conference on Decision and Control, Cancun, 9-11 December 2008, 1358-1363.
 Kaneko, O., Miyachi, M. and Fujii, T. (2011) Simultaneous Updating of Model and Controller Based on Fictitious Reference Iterative Tuning. SICE Journal of Control, Measurement, and System Integration, 4, 63-70.
 Shwe, P.E.E. and Yamamoto, S. (2016) Data-Driven Method to Simultaneously Obtain a Linearized State Space Model and Pole Placement Gain. Proceedings of the 3rd Multi Symposium on Control Systems, Nagoya, 7-10 March 2016, 3B3-2.
 Shwe, P.E.E. and Yamamoto, S. (2016) Real-Time Simultaneously Updating a Linearized State-Space Model and Pole Placement Gain. Proceedings of SICE Annual Conference, Tsukuba, 20-23 September 2016, 196-201.