Multidimensional system

System in which not only one independent variable exists

In mathematical systems theory, a multidimensional system or m-D system is a system in which not only one independent variable exists (like time), but there are several independent variables.

Important problems such as factorization and stability of m-D systems (m > 1) have recently attracted the interest of many researchers and practitioners. The reason is that the factorization and stability is not a straightforward extension of the factorization and stability of 1-D systems because, for example, the fundamental theorem of algebra does not exist in the ring of m-D (m > 1) polynomials.

Applications

Multidimensional systems or m-D systems are the necessary mathematical background for modern digital image processing with many applications in biomedicine, X-ray technology and satellite communications.[1][2] There are also some studies combining m-D systems with partial differential equations (PDEs).

Linear multidimensional state-space model

A state-space model is a representation of a system in which the effect of all "prior" input values is contained by a state vector. In the case of an m-d system, each dimension has a state vector that contains the effect of prior inputs relative to that dimension. The collection of all such dimensional state vectors at a point constitutes the total state vector at the point.

Consider a uniform discrete space linear two-dimensional (2d) system that is space invariant and causal. It can be represented in matrix-vector form as follows:[3][4]

Represent the input vector at each point ( i , j ) {\displaystyle (i,j)} by u ( i , j ) {\displaystyle u(i,j)} , the output vector by y ( i , j ) {\displaystyle y(i,j)} the horizontal state vector by R ( i , j ) {\displaystyle R(i,j)} and the vertical state vector by S ( i , j ) {\displaystyle S(i,j)} . Then the operation at each point is defined by:

R ( i + 1 , j ) = A 1 R ( i , j ) + A 2 S ( i , j ) + B 1 u ( i , j ) S ( i , j + 1 ) = A 3 R ( i , j ) + A 4 S ( i , j ) + B 2 u ( i , j ) y ( i , j ) = C 1 R ( i , j ) + C 2 S ( i , j ) + D u ( i , j ) {\displaystyle {\begin{aligned}R(i+1,j)&=A_{1}R(i,j)+A_{2}S(i,j)+B_{1}u(i,j)\\S(i,j+1)&=A_{3}R(i,j)+A_{4}S(i,j)+B_{2}u(i,j)\\y(i,j)&=C_{1}R(i,j)+C_{2}S(i,j)+Du(i,j)\end{aligned}}}

where A 1 , A 2 , A 3 , A 4 , B 1 , B 2 , C 1 , C 2 {\displaystyle A_{1},A_{2},A_{3},A_{4},B_{1},B_{2},C_{1},C_{2}} and D {\displaystyle D} are matrices of appropriate dimensions.

These equations can be written more compactly by combining the matrices:

[ R ( i + 1 , j ) S ( i , j + 1 ) y ( i , j ) ] = [ A 1 A 2 B 1 A 3 A 4 B 2 C 1 C 2 D ] [ R ( i , j ) S ( i , j ) u ( i , j ) ] {\displaystyle {\begin{bmatrix}R(i+1,j)\\S(i,j+1)\\y(i,j)\end{bmatrix}}={\begin{bmatrix}A_{1}&A_{2}&B_{1}\\A_{3}&A_{4}&B_{2}\\C_{1}&C_{2}&D\end{bmatrix}}{\begin{bmatrix}R(i,j)\\S(i,j)\\u(i,j)\end{bmatrix}}}

Given input vectors u ( i , j ) {\displaystyle u(i,j)} at each point and initial state values, the value of each output vector can be computed by recursively performing the operation above.

Multidimensional transfer function

A discrete linear two-dimensional system is often described by a partial difference equation in the form: p , q = 0 , 0 m , n a p , q y ( i p , j q ) = p , q = 0 , 0 m , n b p , q x ( i p , j q ) {\displaystyle \sum _{p,q=0,0}^{m,n}a_{p,q}y(i-p,j-q)=\sum _{p,q=0,0}^{m,n}b_{p,q}x(i-p,j-q)}

where x ( i , j ) {\displaystyle x(i,j)} is the input and y ( i , j ) {\displaystyle y(i,j)} is the output at point ( i , j ) {\displaystyle (i,j)} and a p , q {\displaystyle a_{p,q}} and b p , q {\displaystyle b_{p,q}} are constant coefficients.

To derive a transfer function for the system the 2d Z-transform is applied to both sides of the equation above.

p , q = 0 , 0 m , n a p , q z 1 p z 2 q Y ( z 1 , z 2 ) = p , q = 0 , 0 m , n b p , q z 1 p z 2 q X ( z 1 , z 2 ) {\displaystyle \sum _{p,q=0,0}^{m,n}a_{p,q}z_{1}^{-p}z_{2}^{-q}Y(z_{1},z_{2})=\sum _{p,q=0,0}^{m,n}b_{p,q}z_{1}^{-p}z_{2}^{-q}X(z_{1},z_{2})}

Transposing yields the transfer function T ( z 1 , z 2 ) {\displaystyle T(z_{1},z_{2})} :

T ( z 1 , z 2 ) = Y ( z 1 , z 2 ) X ( z 1 , z 2 ) = p , q = 0 , 0 m , n b p , q z 1 p z 2 q p , q = 0 , 0 m , n a p , q z 1 p z 2 q {\displaystyle T(z_{1},z_{2})={Y(z_{1},z_{2}) \over X(z_{1},z_{2})}={\sum _{p,q=0,0}^{m,n}b_{p,q}z_{1}^{-p}z_{2}^{-q} \over \sum _{p,q=0,0}^{m,n}a_{p,q}z_{1}^{-p}z_{2}^{-q}}}

So given any pattern of input values, the 2d Z-transform of the pattern is computed and then multiplied by the transfer function T ( z 1 , z 2 ) {\displaystyle T(z_{1},z_{2})} to produce the Z-transform of the system output.

Realization of a 2d transfer function

Often an image processing or other md computational task is described by a transfer function that has certain filtering properties, but it is desired to convert it to state-space form for more direct computation. Such conversion is referred to as realization of the transfer function.

Consider a 2d linear spatially invariant causal system having an input-output relationship described by:

Y ( z 1 , z 2 ) = p , q = 0 , 0 m , n b p , q z 1 p z 2 q p , q = 0 , 0 m , n a p , q z 1 p z 2 q X ( z 1 , z 2 ) {\displaystyle Y(z_{1},z_{2})={\sum _{p,q=0,0}^{m,n}b_{p,q}z_{1}^{-p}z_{2}^{-q} \over \sum _{p,q=0,0}^{m,n}a_{p,q}z_{1}^{-p}z_{2}^{-q}}X(z_{1},z_{2})}

Two cases are individually considered 1) the bottom summation is simply the constant 1 2) the top summation is simply a constant k {\displaystyle k} . Case 1 is often called the "all-zero" or "finite impulse response" case, whereas case 2 is called the "all-pole" or "infinite impulse response" case. The general situation can be implemented as a cascade of the two individual cases. The solution for case 1 is considerably simpler than case 2 and is shown below.

Example: all zero or finite impulse response

Y ( z 1 , z 2 ) = p , q = 0 , 0 m , n b p , q z 1 p z 2 q X ( z 1 , z 2 ) {\displaystyle Y(z_{1},z_{2})=\sum _{p,q=0,0}^{m,n}b_{p,q}z_{1}^{-p}z_{2}^{-q}X(z_{1},z_{2})}

The state-space vectors will have the following dimensions:

R ( 1 × m ) , S ( 1 × n ) , x ( 1 × 1 ) {\displaystyle R(1\times m),\quad S(1\times n),\quad x(1\times 1)} and y ( 1 × 1 ) {\displaystyle y(1\times 1)}

Each term in the summation involves a negative (or zero) power of z 1 {\displaystyle z_{1}} and of z 2 {\displaystyle z_{2}} which correspond to a delay (or shift) along the respective dimension of the input x ( i , j ) {\displaystyle x(i,j)} . This delay can be effected by placing 1 {\displaystyle 1} ’s along the super diagonal in the A 1 {\displaystyle A_{1}} . and A 4 {\displaystyle A_{4}} matrices and the multiplying coefficients b i , j {\displaystyle b_{i,j}} in the proper positions in the A 2 {\displaystyle A_{2}} . The value b 0 , 0 {\displaystyle b_{0,0}} is placed in the upper position of the B 1 {\displaystyle B_{1}} matrix, which will multiply the input x ( i , j ) {\displaystyle x(i,j)} and add it to the first component of the R i , j {\displaystyle R_{i,j}} vector. Also, a value of b 0 , 0 {\displaystyle b_{0,0}} is placed in the D {\displaystyle D} matrix which will multiply the input x ( i , j ) {\displaystyle x(i,j)} and add it to the output y {\displaystyle y} . The matrices then appear as follows:

A 1 = [ 0 0 0 0 0 1 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 1 0 ] {\displaystyle A_{1}={\begin{bmatrix}0&0&0&\cdots &0&0\\1&0&0&\cdots &0&0\\0&1&0&\cdots &0&0\\\vdots &\vdots &\vdots &\ddots &\vdots &\vdots \\0&0&0&\cdots &0&0\\0&0&0&\cdots &1&0\end{bmatrix}}}
A 2 = [ 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 ] {\displaystyle A_{2}={\begin{bmatrix}0&0&0&\cdots &0&0\\0&0&0&\cdots &0&0\\0&0&0&\cdots &0&0\\\vdots &\vdots &\vdots &\ddots &\vdots &\vdots \\0&0&0&\cdots &0&0\\0&0&0&\cdots &0&0\end{bmatrix}}}
A 3 = [ b 1 , n b 2 , n b 3 , n b m 1 , n b m , n b 1 , n 1 b 2 , n 1 b 3 , n 1 b m 1 , n 1 b m , n 1 b 1 , n 2 b 2 , n 2 b 3 , n 2 b m 1 , n 2 b m , n 2 b 1 , 2 b 2 , 2 b 3 , 2 b m 1 , 2 b m , 2 b 1 , 1 b 2 , 1 b 3 , 1 b m 1 , 1 b m , 1 ] {\displaystyle A_{3}={\begin{bmatrix}b_{1,n}&b_{2,n}&b_{3,n}&\cdots &b_{m-1,n}&b_{m,n}\\b_{1,n-1}&b_{2,n-1}&b_{3,n-1}&\cdots &b_{m-1,n-1}&b_{m,n-1}\\b_{1,n-2}&b_{2,n-2}&b_{3,n-2}&\cdots &b_{m-1,n-2}&b_{m,n-2}\\\vdots &\vdots &\vdots &\ddots &\vdots &\vdots \\b_{1,2}&b_{2,2}&b_{3,2}&\cdots &b_{m-1,2}&b_{m,2}\\b_{1,1}&b_{2,1}&b_{3,1}&\cdots &b_{m-1,1}&b_{m,1}\end{bmatrix}}}

A 4 = [ 0 0 0 0 0 1 0 0 0 0 0 1 0 0 0 0 0 0 0 0 0 0 0 1 0 ] {\displaystyle A_{4}={\begin{bmatrix}0&0&0&\cdots &0&0\\1&0&0&\cdots &0&0\\0&1&0&\cdots &0&0\\\vdots &\vdots &\vdots &\ddots &\vdots &\vdots \\0&0&0&\cdots &0&0\\0&0&0&\cdots &1&0\end{bmatrix}}}

B 1 = [ 1 0 0 0 0 0 ] {\displaystyle B_{1}={\begin{bmatrix}1\\0\\0\\0\\\vdots \\0\\0\end{bmatrix}}}
B 2 = [ b 0 , n b 0 , n 1 b 0 , n 2 b 0 , 2 b 0 , 1 ] {\displaystyle B_{2}={\begin{bmatrix}b_{0,n}\\b_{0,n-1}\\b_{0,n-2}\\\vdots \\b_{0,2}\\b_{0,1}\end{bmatrix}}}
C 1 = [ b 1 , 0 b 2 , 0 b 3 , 0 b m 1 , 0 b m , 0 ] {\displaystyle C_{1}={\begin{bmatrix}b_{1,0}&b_{2,0}&b_{3,0}&\cdots &b_{m-1,0}&b_{m,0}\\\end{bmatrix}}}
C 2 = [ 0 0 0 0 1 ] {\displaystyle C_{2}={\begin{bmatrix}0&0&0&\cdots &0&1\\\end{bmatrix}}}
D = [ b 0 , 0 ] {\displaystyle D={\begin{bmatrix}b_{0,0}\end{bmatrix}}}

[3][4]

References

  1. ^ Bose, N.K., ed. (1985). Multidimensional Systems Theory, Progress, Directions and Open Problems in Multidimensional Systems. Dordre http, Holland: D. Reidel Publishing Company.
  2. ^ Bose, N.K., ed. (1979). Multidimensional Systems: Theory and Applications. IEEE Press.
  3. ^ a b Tzafestas, S.G., ed. (1986). Multidimensional Systems: Techniques and Applications. New York: Marcel-Dekker.
  4. ^ a b Kaczorek, T. (1985). Two-Dimensional Linear Systems. Lecture Notes Contr. and Inform. Sciences. Vol. 68. Springer-Verlag.