+ All Categories
Home > Documents > 08 Integral Action

08 Integral Action

Date post: 05-Apr-2018
Category:
Upload: noori734
View: 222 times
Download: 0 times
Share this document with a friend

of 15

Transcript
  • 8/2/2019 08 Integral Action

    1/15

  • 8/2/2019 08 Integral Action

    2/15

    Lecture: Integral action in state feedback control Adjustment of DC-gain for reference tracking

    Reference tracking

    Assume the open-loop system completely reachable and observable

    We know state feedback we can bring the output y(k) to zero asymptoticallyHow to make the output y(k) track a generic constant set-point r(k) r ?

    Solution: set u(k) = Kx(k) +v(k)

    v(k) = Fr(k)

    We need to choose gain F properly to ensure reference tracking

    !"#$%&'$()*+,'-..

    r(k) u(k) x(k) y(k)

    ++

    ',#/+,((-+

    v(k)

    x(k + 1) = (A +BK)x(k) +BFr(k)

    y(k) = Cx(k)

    Prof. Alberto Bemporad (University of Trento) Automatic Control 1 Academic year 2010-2011 2 / 15

  • 8/2/2019 08 Integral Action

    3/15

    Lecture: Integral action in state feedback control Adjustment of DC-gain for reference tracking

    Reference tracking

    To have y(k) r we need a unit DC-gain from r to y

    C(I (A +BK))1BF= I

    Assume we have as many inputs as outputs (example: u,y )

    Assume the DC-gain from u to y is invertible, that is CAdj(IA)B invertible

    Since state feedback doesnt change the zeros in closed-loop

    CAdj(IABK)B = CAdj(IA)B

    then CAdj(IABK)B is also invertibleSet

    F= (C(I (A +BK))1B)1

    Prof. Alberto Bemporad (University of Trento) Automatic Control 1 Academic year 2010-2011 3 / 15

  • 8/2/2019 08 Integral Action

    4/15

    L I l i i f db k l Adj f DC i f f ki

  • 8/2/2019 08 Integral Action

    5/15

    Lecture: Integral action in state feedback control Adjustment of DC-gain for reference tracking

    Reference tracking

    Problem: we have no direct feedback on the tracking error e(k) = y(k) r(k)

    Will this solution be robust with respect to model uncertainties and

    exogenous disturbances ?

    Consider an input disturbance d(k) (modeling for instance a non-ideal

    actuator, or an unmeasurable disturbance)

    !"#$%&'$()*+,'-..

    r(k) u(k) x(k) y(k)d(k)

    +

    + +

    +

    */0)!&.0/+1$#'-

    ',#0+,((-+

    Prof. Alberto Bemporad (University of Trento) Automatic Control 1 Academic year 2010-2011 5 / 15

    L t I t l ti i t t f db k t l Adj t t f DC i f f t ki

  • 8/2/2019 08 Integral Action

    6/15

    Lecture: Integral action in state feedback control Adjustment of DC-gain for reference tracking

    Example (contd)

    Let the input disturbance d(k) = 0.01, k = 0,1,.. .

    0 10 20 30 400

    0.2

    0.4

    0.6

    0.8

    1

    1.2

    1.4

    sample steps

    The reference is not tracked !

    The unmeasurable disturbance d(k) has modified the nominal conditions for

    which we designed our controller

    Prof. Alberto Bemporad (University of Trento) Automatic Control 1 Academic year 2010-2011 6 / 15

    Lecture: Integral action in state feedback control Integral action

  • 8/2/2019 08 Integral Action

    7/15

    Lecture: Integral action in state feedback control Integral action

    Integral action for disturbance rejection

    Consider the problem of regulating the output y(k) to r(k) 0 under the

    action of the input disturbance d(k)Lets augment the open-loop system with the integral of the output vector

    q(k + 1) = q(k) +y(k)

    integral action

    The augmented system isx(k + 1)

    q(k + 1)

    =

    A 0

    C I

    x(k)

    q(k)

    +

    B

    0

    u(k) +

    B

    0

    d(k)

    y(k) = C 0 x(k)q(k) Design a stabilizing feedback controller for the augmented system

    u(k) =

    K H

    x(k)

    q(k)

    Prof. Alberto Bemporad (University of Trento) Automatic Control 1 Academic year 2010-2011 7 / 15

    Lecture: Integral action in state feedback control Integral action

  • 8/2/2019 08 Integral Action

    8/15

    Lecture: Integral action in state feedback control Integral action

    Rejection of constant disturbances

    !"#$%&'$()*+,'-..

    +

    + +

    +

    */0)!&.0/+1$#'-

    -2+$()$'0&,#

    .0$0-)3--!1$'4

    q(k)

    u(k)d(k)

    x(k) y(k)

    Theorem

    Assume a stabilizing gain [H K] can be designed for the system

    augmented with integral action. Then limk+y(k) = 0 for all constant

    disturbances d(k) d

    Prof. Alberto Bemporad (University of Trento) Automatic Control 1 Academic year 2010-2011 8 / 15

    Lecture: Integral action in state feedback control Integral action

  • 8/2/2019 08 Integral Action

    9/15

    Lecture: Integral action in state feedback control Integral action

    Rejection of constant disturbances

    !"#$%&'$()*+,'-..

    +

    + +

    +

    */0)!&.0/+1$#'-

    -2+$()$'0&,#

    .0$0-)3--!1$'4

    q(k)

    u(k)

    d(k)

    x(k) y(k)

    Proof:

    The state-update matrix of the closed-loop system is

    A 0

    C I+B

    0 K HThe matrix has asymptotically stable eigenvalues by construction

    For a constant excitation d(k) the extended state

    x(k)q(k)

    converges to a

    steady-state value, in particular limk q(k) = q

    Hence, limky(k) = limk q(k + 1)

    q(k) = q

    q = 0

    Prof. Alberto Bemporad (University of Trento) Automatic Control 1 Academic year 2010-2011 9 / 15

    Lecture: Integral action in state feedback control Integral action

  • 8/2/2019 08 Integral Action

    10/15

    g g

    Example (contd) Now with integral action

    Poles placed in (0.8 0.2j,0.3) for the augmented system. Resulting closed-loop:

    x(k + 1) = 1.1 10 0.8

    x(k) + 01 (u(k) + d(k))

    q(k + 1) = q(k) +y(k)

    y(k) =

    1 0

    x(k)

    u(k) = 0.48 1 x(k) 0.056q(k)Closed-loop simulation for x(0) = [0 0], d(k) 1:

    0 10 20 30 400.5

    0

    0.5

    1

    1.5

    2

    2.5

    3

    sample steps

    Prof. Alberto Bemporad (University of Trento) Automatic Control 1 Academic year 2010-2011 10 / 15

    Lecture: Integral action in state feedback control Integral action

  • 8/2/2019 08 Integral Action

    11/15

    g g

    Integral action for set-point tracking

    !"#$%&'$()*+,'-..

    +

    + +

    +

    */0)!&.0/+1$#'-

    -2+$()

    $'0&,#

    .0$0-)3--!1$'4

    q(k) u(k)d(k)

    x(k) y(k)

    -

    +r(k)

    0+$'4)-++,+

    Idea: Use the same feedback gains (K,H) designed earlier, but instead of feeding

    back the integral of the output, feed back the integral of the tracking error

    q(k + 1) = q(k) + (y(k) r(k)) integral action

    Prof. Alberto Bemporad (University of Trento) Automatic Control 1 Academic year 2010-2011 11 / 15

    Lecture: Integral action in state feedback control Integral action

  • 8/2/2019 08 Integral Action

    12/15

    Example (contd)

    x(k + 1) =

    1.1 1

    0 0.8

    x(k)

    +

    0

    1 (u(k) + d(k))

    q(k + 1) = q(k) + (y(k) r(k)) tracking error

    y(k) =

    1 0

    x(k)

    u(k) = 0.48 1 x(k) 0.056q(k)0 10 20 30 40

    0

    0.5

    1

    1.5

    2

    2.5

    3

    3.5

    sample steps

    Response for x(0) = [0 0]

    ,d(k) 1, r(k) 1

    Looks like its working . . . but why ?

    Prof. Alberto Bemporad (University of Trento) Automatic Control 1 Academic year 2010-2011 12 / 15

    Lecture: Integral action in state feedback control Integral action

  • 8/2/2019 08 Integral Action

    13/15

    Tracking & rejection of constant disturbances/set-points

    Theorem

    Assume a stabilizing gain [H K] can be designed for the systemaugmented with integral action. Then limk+y(k) = r for all constant

    disturbances d(k) d and set-points r(k) r

    Proof:

    The closed-loop systemx(k + 1)

    q(k + 1)

    =

    A +BK BH

    C I

    x(k)

    q(k)

    +

    B 0

    0 I

    d(k)

    r(k)

    y(k) = C 0 x(k)q(k) has input

    d(k)r(k)

    and is asymptotically stable by construction

    For a constant excitation

    d(k)r(k)

    the extended state

    x(k)q(k)

    converges to a

    steady-state value, in particular limk q(k) = q

    Hence, limky(k) r(k) = limk q(k + 1) q(k) = q q = 0 Prof. Alberto Bemporad (University of Trento) Automatic Control 1 Academic year 2010-2011 13 / 15

    Lecture: Integral action in state feedback control Integral action

  • 8/2/2019 08 Integral Action

    14/15

    Integral action for continuous-time systems

    The same reasoning can be applied to continuous-time systems

    x(t) = Ax(t) +Bu(t)y(t) = Cx(t)

    Augment the system with the integral of the output q(t) =t

    0y()d, i.e.,

    q(t) = y(t) = Cx(t)

    integral actionThe augmented system is

    d

    dt

    x(t)

    q(t)

    =

    A 0

    C 0

    x(t)

    q(t)

    +

    B

    0

    u(t)

    y(t) = C 0 x(t)q(t) Design a stabilizing controller [K H] for the augmented system

    Implement

    u(t) = Kx(t) +Ht

    0

    (y() r())d

    Prof. Alberto Bemporad (University of Trento) Automatic Control 1 Academic year 2010-2011 14 / 15

    Lecture: Integral action in state feedback control Integral action

  • 8/2/2019 08 Integral Action

    15/15

    English-Italian Vocabulary

    reference tracking inseguimento del riferimento

    steady state regime stazionario

    set point livello di riferimento

    Translation is obvious otherwise.

    Prof. Alberto Bemporad (University of Trento) Automatic Control 1 Academic year 2010-2011 15 / 15


Recommended