+ All Categories
Home > Documents > Challenges in Evaluating a Prototype Project in a Large ... › fr › system › files ›...

Challenges in Evaluating a Prototype Project in a Large ... › fr › system › files ›...

Date post: 28-Jan-2021
Category:
Upload: others
View: 0 times
Download: 0 times
Share this document with a friend
18
Research and Practice Notes/Notes sur la recherche et les méthodes Challenges in Evaluating a Prototype Project in a Large Health Authority: Lessons Learned M. Elizabeth Snow Centre for Health Evaluation and Outcome Sciences (CHÉOS) Vancouver, BC Michelle Urbina-Beggs Fraser Health Authority Maple Ridge, BC Tamara Van Tent Fraser Health Authority Abbotsford, BC Abstract: A developmental evaluation was undertaken to evaluate a prototype test of a new model of perinatal healthcare across acute maternity, public health, and primary care in two hospitals in a large health authority. e project was initiated to bridge gaps in care across the acute and community settings to ensure a seamless per- inatal healthcare journey for women. e objective of the evaluation was to support the prototyping process by providing data to inform decisions as the prototype was developed and by documenting decisions as they were made. is article explores challenges faced during the evaluation, including unfamiliarity of the health sector with prototype projects and their inherent uncertainty, a disconnect between the rapid pace of a prototype project and bureaucratic hurdles of working within a large organization, and high leadership turnover throughout the project. How these chal- lenges were addressed, and the lessons learned for future evaluations, are discussed. Keywords: developmental evaluation, healthcare, interdisciplinarity, prototyping Résumé : Une évaluation évolutive a été entreprise pour évaluer un essai de proto- type d’un nouveau modèle de prestation de soins de santé périnataux déployé dans les unités de soins de maternité de courte durée, de soins de santé publique et de soins de santé primaires de deux hôpitaux d’une grande régie de la santé. Le projet a été lancé pour combler les lacunes dans les soins à travers les contextes aigus et com- munautaires pour s’assurer que les expériences des femmes dans les soins de santé périnatale sont homogènes. L’objectif de l’évaluation était de soutenir le processus de prototypage en fournissant des données pour aider à éclairer les décisions tout au long du développement du prototype et documenter les décisions prises. Cet article © 2015 Canadian Journal of Program Evaluation / La Revue canadienne d'évaluation de programme 30.2 (Fall / automne), 177–194 doi: 10.3138/cjpe.209 Corresponding author: M. Elizabeth Snow, [email protected]
Transcript
  • Research and Practice Notes/Notes sur la recherche et les méthodes

    Challenges in Evaluating a Prototype Project in a Large Health Authority: Lessons Learned

    M. Elizabeth Snow Centre for Health Evaluation and Outcome Sciences (CHÉOS)

    Vancouver, BC

    Michelle Urbina-Beggs Fraser Health Authority

    Maple Ridge, BC

    Tamara Van Tent Fraser Health Authority

    Abbotsford, BC

    Abstract: A developmental evaluation was undertaken to evaluate a prototype test of a new model of perinatal healthcare across acute maternity, public health, and primary care in two hospitals in a large health authority. Th e project was initiated to bridge gaps in care across the acute and community settings to ensure a seamless per-inatal healthcare journey for women. Th e objective of the evaluation was to support the prototyping process by providing data to inform decisions as the prototype was developed and by documenting decisions as they were made. Th is article explores challenges faced during the evaluation, including unfamiliarity of the health sector with prototype projects and their inherent uncertainty, a disconnect between the rapid pace of a prototype project and bureaucratic hurdles of working within a large organization, and high leadership turnover throughout the project. How these chal-lenges were addressed, and the lessons learned for future evaluations, are discussed.

    Keywords: developmental evaluation, healthcare, interdisciplinarity, prototyping

    Résumé : Une évaluation évolutive a été entreprise pour évaluer un essai de proto-type d’un nouveau modèle de prestation de soins de santé périnataux déployé dans les unités de soins de maternité de courte durée, de soins de santé publique et de soins de santé primaires de deux hôpitaux d’une grande régie de la santé. Le projet a été lancé pour combler les lacunes dans les soins à travers les contextes aigus et com-munautaires pour s’assurer que les expériences des femmes dans les soins de santé périnatale sont homogènes. L’objectif de l’évaluation était de soutenir le processus de prototypage en fournissant des données pour aider à éclairer les décisions tout au long du développement du prototype et documenter les décisions prises. Cet article

    © 2015 Canadian Journal of Program Evaluation / La Revue canadienne d'évaluation de programme30.2 (Fall / automne), 177–194 doi: 10.3138/cjpe.209

    Corresponding author : M. Elizabeth Snow, [email protected]

    mailto:[email protected]

  • 178 Snow, Urbina-Beggs, and Van Tent

    © 2015 CJPE 30.2, 177–194 doi: 10.3138/cjpe.209

    explore les défi s rencontrés durant l’évaluation, notamment le manque de familiarité du secteur de la santé avec les projets de mise à l’essai de prototypes et leur incertitude inhérente, le fossé entre le rythme rapide de tels projets et les obstacles bureaucra-tiques liés au travail au sein d’une grande organisation ainsi qu’un haut taux de roulement des personnes responsables de la direction du projet. La manière dont ces défi s ont été gérés ainsi que les leçons apprises, lesquelles éclaireront les futures évaluations, sont aussi abordées.

    Mots clés : évaluation évolutive, soins de santé, interdisciplinarité, prototypage

    DESCRIPTION OF CASE AND EVALUATION CONTEXT

    Why Was the Evaluation Conducted? What Did the Client Want to Learn? Gaps in care at transition points between public health, primary healthcare, acute maternity, and neonatal intensive care units (NICUs) in a large, regional health authority created duplication, ineffi ciencies, and reduced quality of care. A need to address specifi c care gaps (e.g., inadequate information sharing among healthcare providers as clients move between hospital and community settings) was recognized based on fi ndings from a literature review, gap analysis, and healthcare provider and client surveys ( Chopova, 2011 ; Fraser Health, 2011 ). A cross-disciplinary steering committee (SC) and planning team were created to develop a new collaborative model for providing perinatal care to bridge the gaps. Th e model focused on coordinating care for high-risk and vulnerable clients across transitions between hospital and community settings by promoting seam-less communication, coordination, and collaboration (the “3 Cs”) between all in-volved healthcare providers for the best care of women, babies, and families. Th is compelling vision resonated with care providers from all disciplines and served as a touchstone for all involved in the project to help guide decision-making during the prototyping process.

    Th e initial conceptual model, called the Seamless Perinatal Transition Team ( Figure 1 ), listed the roles included on the team, and indicated that a referral form is used to refer clients to the team and that feedback should be provided to the referring party; however, the model does not detail the ways in which team mem-bers interact with one another or with clients. To support the transition team’s work, a referral form was created and distributed to potential referral sources (e.g., public health unit, primary maternity care providers), and a feedback form was created for the transition teams to close the communications loop by informing the referring party of their referral’s outcome. Th e model was prototyped from November 2012 to December 2013 ( Figure 2 ). As it was intended to use the fi nd-ings from the prototyping to inform the roll-out of the model at all maternity hospitals in the health authority, the prototype was tested in one large tertiary hospital and one smaller community hospital, to determine which elements would work in these diff erent contexts.

  • Evaluating a Prototype Project 179

    CJPE 30.2, 177–194 © 2015doi: 10.3138/cjpe.209

    Th e process of prototyping, which is common in computer sciences and engineering, is gaining traction in healthcare ( Ferguson, n.d. ). As opposed to a pilot project, in which a model is fully designed and then launched at a test site, a prototype project involves part of the design work being conducted on the ground by the end users—in this case, front-line healthcare staff —where ideas can be tested in an iterative fashion in a real-world setting. In a pilot project, a model including all the details of the protocols and procedures would be created; the evaluation would typically involve assessing if the pilot was implemented as designed and, if so, whether or not the intended outcomes were achieved. Pro-totyping also diff ers from continuous quality improvement, a methodology used in healthcare that focuses on improving existing processes through incremental improvements ( American Society for Quality, n.d. ), while prototyping is used to design entirely new processes.

    Figure 1. Seamless Perinatal Transition Team Model

    Client accesses care(e.g., through primary care provider or public

    health)

    Referral to Seamless Perinatal Transi�on Team

    referral form

    Public health nurse

    Perinatal inpa�ent

    nurse

    Antenatal outpa�ent

    nurse

    Hospital social worker

    Neonatal intensive care unit

    nurse

    Complex Care Plan

    feed

    back

    form

    Guiding Principles: Women- and Family-Centred Care – Collabora�on – Quality and Safety

  • 180 Snow, Urbina-Beggs, and Van Tent

    © 2015 CJPE 30.2, 177–194 doi: 10.3138/cjpe.209

    Figu

    re 2

    . Pro

    ject

    Tim

    elin

    e

    Ini�

    al m

    odel

    de

    velo

    pmen

    tPr

    otot

    ypin

    gRo

    ll ou

    t to

    othe

    r ho

    spita

    ls

    deve

    lopm

    enta

    l ev

    alua

    �on

    June

    Au

    gO

    ctDe

    cFe

    b

    A

    pr

    J

    unAu

    g

    Oct

    Dec

    Feb

    2012

    2012

    20

    1220

    1220

    13

    201

    3

    201

    3

    2

    013

    201

    3

    2

    013

    20

    14

    Impl

    emen

    ta�o

    n Pl

    anni

    ng T

    eam

    mee

    �ng

    Stee

    ring

    Com

    mi�

    ee m

    ee�n

    g

    Advi

    sory

    Com

    mi�

    ee m

    ee�n

    g*

    Lege

    nd

    *Adv

    isory

    Com

    mi�

    ee m

    embe

    rs a

    lso

    cons

    ulte

    d vi

    a ph

    one/

    emai

    l on

    an a

    d ho

    c ba

    sis

  • Evaluating a Prototype Project 181

    CJPE 30.2, 177–194 © 2015doi: 10.3138/cjpe.209

    In this project, the prototyping process required front-line teams to op-erationalize the model, experimenting to determine what worked in each site as diff erent situations arose. For example, the initial model showed that dif-ferent types of healthcare providers (who previously worked independently of one another in separate settings) were intended to work together to provide client-centred care in the maternity unit, but it did not identify in what ways they would interact with one another and with the client; those details were determined by front-line providers during the prototyping process, with a focus on testing options and learning from those tests. Details worked out dur-ing prototyping included whether public health nurses (PHNs), who had not previously been working in the hospital, would interact directly with clients or only work with the other healthcare providers; how oft en PHNs would attend the maternity unit and for how long; and how acute care nurses would update PHNs on client issues occurring when PHNs were not onsite at the hospital. Also, referral and feedback forms were substantially changed by front-line nurses as they learned what information was required—and what information was not required—to do their work. Management, the front-line staff engaged in the prototyping, and the evaluator met regularly to discuss their experiences as the implementation planning team (IPT); regular IPT meetings allowed staff from both sites to share learnings with one another. For decisions beyond this group’s scope, recommendations were brought to the SC, comprising manage-ment with decision-making authority for the project. An advisory committee (AC) provided advice to the SC ( Table 1 ).

    A developmental evaluation (DE) approach ( Patton, 2011 ) was chosen to allow the evaluation to be adaptive and responsive to the prototype nature of the project, where fi nal details of the model emerged from the implementation. Table 2 lists the characteristics of a situation that makes it suited to DE as de-scribed by Dozois, Langlois, & Blanchet-Cohen (2010) .

    Th e objective of the evaluation was to support the Seamless Perinatal project through tracking the model’s development throughout the prototyping process, providing data to inform decisions, and documenting decisions as they were made. Specifi cally, the evaluation addressed the following questions:

    • What have we learned from implementing this prototype? ° What worked well? ° What challenges were encountered and how were they dealt with? ° What could be improved? How could it be improved?

    • What has the prototype taught us that can be applied to spreading Seam-less Perinatal Transition Teams to other maternity hospitals?

    In addition, three “intended contribution statements,” which are measurable progress markers providing stakeholders with goals to orient them in their work, were set for the project ( Table 3 ).

  • 182 Snow, Urbina-Beggs, and Van Tent

    © 2015 CJPE 30.2, 177–194 doi: 10.3138/cjpe.209

    Table 1. Governance Committee Membership

    Governance committee Membership

    Implementation planning team (IPT) • Antepartum outpatient nurses• Public health nurses• Maternity unit managers• Public health unit supervisors• Public health unit managers• Physician• Social work leads• Project manager• Evaluation specialist

    Steering committee (SC) • Director, public health• Director, acute maternity• Medical lead• Clinical nurse specialist• Public health unit managers• Maternity managers• Antenatal outpatient nurse• Physician• Project manager• Evaluation specialist

    Advisory committee (AC) • Director, public health• Director, acute maternity• Director, maternity/neonatal intensive

    care• Director, maternity/child/youth• Clinical nurse specialist• Program medical director• Medical health offi cers• Decision support services lead• Health information management leads• Privacy lead• Perinatal outpatient clinic lead• Quality improvement/patient safety

    consultants• Project manager• Evaluation specialist

  • Evaluating a Prototype Project 183

    CJPE 30.2, 177–194 © 2015doi: 10.3138/cjpe.209

    Table 2. Elements That Make a Situation Appropriate for Developmental Evaluation

    Elements that make a situation appropriate for developmental evaluation

    How it applied to the project

    Highly emergent and volatile (e.g., the environment is always changing)

    • Prototype nature of the project meant the design of the model was emergent from the work on the front-line.

    • Healthcare settings are inherently complex (Begun, Zimmerman, & Dooley, 2003).

    Diffi cult to plan or predict be-cause variables are interdepend-ent and nonlinear

    • The operationalization of the model emerged from interactions between an interdepend-ent group of healthcare practitioners, making results diffi cult to plan and predict.

    Socially complex, requiring col-laboration among stakeholders from diff erent organizations, systems, and/or sectors

    • Bringing together acute and community healthcare providers—who have diff erent priorities, languages, and cultures—to work in close collaboration was socially complex.

    • Multiple levels within the organization (from front lines to senior leadership) were required to work together, adding another layer of complexity.

    Innovative, requiring real-time learning and development

    • Prototype nature of the project required re-al-time learning as the model was developed by front-line staff .

    What resources (time, money, in-kind, etc.) were available for conducting the evaluation? Were they suitable for answering the evaluation questions? Th e internal evaluator from the Population & Public Health program was ap-proached to conduct the evaluation work for the prototype. Given the DE approach taken, the evaluator was embedded at all levels of the project govern-ance structure as a member of the AC, SC, and IPT.

    Th e DE process involved tracking the prototype model implementation at both sites, facilitating transparent decision-making, and generating data to assess the model, thus supporting accountability while allowing the fl exibility necessary for an emergent process ( Dozois et al., 2010 ; Rey et al., 2014 ). Examples of types of data collected and what they were used for is provided in Table 4 . Data were brought to meetings of all levels of the governance structure to be collectively in-terpreted and to determine how to apply learnings from the data to further model testing and development ( Rey et al., 2014 ).

  • 184 Snow, Urbina-Beggs, and Van Tent

    © 2015 CJPE 30.2, 177–194 doi: 10.3138/cjpe.209

    Tabl

    e 3.

    The

    Evo

    lutio

    n of

    the

    Inte

    nded

    Con

    trib

    utio

    ns fo

    r the

    Pro

    ject

    Dom

    ain

    Ori

    gina

    l int

    ende

    d co

    n-tr

    ibut

    ion

    stat

    emen

    tW

    hat w

    as le

    arne

    dA

    ctio

    n ta

    ken

    as a

    resu

    lt of

    le

    arni

    ngN

    ew in

    tend

    ed

    cont

    ribu

    tion

    stat

    emen

    t

    Com

    mun

    ica-

    tion

    100%

    of r

    efer

    rals

    will

    re

    ceiv

    e a

    feed

    back

    fo

    rm.

    • M

    ost o

    f the

    fi el

    ds o

    n th

    e or

    igin

    al fe

    edba

    ck

    form

    repe

    ated

    wha

    t was

    on

    the

    refe

    rral

    fo

    rm, r

    esul

    ting

    in d

    uplic

    atio

    n of

    eff o

    rt to

    fi ll

    it ou

    t onl

    y to

    pro

    vide

    the

    refe

    rrin

    g pa

    rty

    with

    info

    rmat

    ion

    they

    alre

    ady

    knew

    .

    An

    inte

    rdis

    cipl

    inar

    y Sh

    ared

    W

    ork

    Team

    cre

    ated

    a p

    roto

    col

    for c

    omm

    unic

    atio

    n am

    ong

    the

    disc

    iplin

    es w

    ith re

    spec

    t to

    perin

    atal

    clie

    nts,

    rath

    er th

    an

    requ

    iring

    that

    a fe

    edba

    ck fo

    rm

    be u

    sed

    as th

    e m

    etho

    d of

    com

    -m

    unic

    atin

    g fr

    om th

    e ho

    spita

    l m

    ater

    nity

    uni

    t to

    com

    mun

    ity-

    base

    d he

    alth

    care

    pro

    vide

    rs.

    Doc

    umen

    ted

    prot

    ocol

    for

    com

    mun

    ica-

    tion

    amon

    g th

    e di

    scip

    lines

    w

    ith re

    spec

    t to

    perin

    atal

    clie

    nts

    is c

    reat

    ed.

    • Th

    e la

    rge

    tert

    iary

    hos

    pita

    l had

    mor

    e cl

    ient

    s re

    ferr

    ed to

    the

    Tran

    sitio

    n Te

    am a

    fter

    they

    ha

    d gi

    ven

    birt

    h in

    the

    hosp

    ital t

    han

    wer

    e re

    ferr

    ed p

    rena

    tally

    , so

    ther

    e w

    as n

    ot a

    co

    mm

    unity

    -bas

    ed re

    ferr

    ing

    part

    y to

    “fee

    d ba

    ck” i

    nfor

    mat

    ion

    to. I

    nste

    ad, t

    he tr

    ansi

    tion

    team

    nee

    ded

    to in

    form

    prim

    ary

    care

    pro

    -vi

    der (

    and

    othe

    r rel

    evan

    t pro

    vide

    rs) a

    bout

    is

    sues

    rela

    ted

    to c

    lient

    .Co

    llabo

    ratio

    n10

    0% o

    f wom

    en w

    ho

    need

    com

    plex

    car

    e fro

    m

    at le

    ast 3

    of t

    he p

    artn

    ers

    on th

    e Tr

    ansi

    tion

    Team

    (i.

    e., p

    rimar

    y ca

    re p

    ro-

    vide

    r, PH

    , acu

    te m

    ater

    -ni

    ty, N

    ICU

    , soc

    ial w

    ork)

    re

    ceiv

    e a

    care

    pla

    n.

    No

    com

    plex

    car

    e pl

    ans

    wer

    e co

    mpl

    eted

    , so

    gro

    up w

    as p

    rom

    pted

    to re

    -exa

    min

    e th

    e de

    fi niti

    on o

    f “co

    mpl

    ex c

    are

    plan

    .” Aft

    er d

    ecid

    -in

    g it

    did

    not fi

    t w

    ith th

    eir w

    ork.

    , th

    ey in

    stea

    d re

    -defi

    ned

    it a

    s “co

    llabo

    rativ

    e ca

    re p

    lan.

    A p

    roce

    ss w

    as d

    evel

    oped

    to

    crea

    te “c

    olla

    bora

    tive

    care

    pla

    ns.”

    Proc

    ess

    is

    deve

    lope

    d fo

    r cr

    eatin

    g co

    l-la

    bora

    tive

    care

    pl

    ans.

  • Evaluating a Prototype Project 185

    CJPE 30.2, 177–194 © 2015doi: 10.3138/cjpe.209

    Coor

    dina

    tion

    100%

    of h

    ealth

    care

    pr

    ovid

    ers

    on th

    e Pe

    ri-na

    tal T

    rans

    ition

    Team

    s re

    port

    und

    erst

    andi

    ng

    of ro

    les

    of a

    ll m

    embe

    rs

    of th

    e te

    am.

    Whe

    n an

    att

    empt

    was

    mad

    e to

    mea

    sure

    the

    exte

    nt to

    whi

    ch th

    e va

    rious

    type

    s of

    hea

    lth-

    care

    pro

    vide

    rs u

    nder

    stoo

    d ea

    ch o

    ther

    s’ ro

    les,

    it w

    as d

    isco

    vere

    d th

    at th

    ere

    was

    no

    expl

    icit

    artic

    ulat

    ion

    of w

    hat t

    hose

    role

    s w

    ere.

    Each

    gro

    up w

    as ta

    sked

    with

    pr

    ovid

    ing

    a do

    cum

    ente

    d de

    scrip

    tion

    of th

    eir d

    isci

    plin

    e’s

    role

    on

    the

    Seam

    less

    Per

    inat

    al

    Tran

    sitio

    n Te

    am. T

    his a

    ctiv

    ity

    help

    ed to

    elim

    inat

    ion

    conf

    usio

    n ar

    ound

    role

    s, re

    duce

    d du

    plic

    a-tio

    n, h

    ighl

    ight

    ed th

    e di

    scip

    lines

    ’ sh

    ared

    goa

    l of p

    rovi

    ding

    the

    best

    pos

    sibl

    e ca

    re to

    the

    patie

    nt,

    and

    info

    rmed

    the

    way

    s in

    whi

    ch

    the

    disc

    iplin

    es w

    orke

    d to

    geth

    er

    to su

    ppor

    t the

    ir pa

    tient

    s.

    Doc

    umen

    ted

    desc

    riptio

    ns

    of th

    e ro

    le o

    f ea

    ch m

    embe

    r of

    the

    Seam

    less

    Pe

    rinat

    al T

    rans

    i-tio

    n Te

    am a

    re

    crea

    ted.

  • 186 Snow, Urbina-Beggs, and Van Tent

    © 2015 CJPE 30.2, 177–194 doi: 10.3138/cjpe.209

    Table 4. Data Collection

    Type of data collected How it was used

    Data from referral forms, including numbers of referrals and charac-teristics of the clients being referred

    • Provided a picture of who the clients being served by the Transition Teams were, including their demographic profi le and risk factors

    • Provided a picture of the workload

    Details on feedback being provided to referral party

    • Provided data on how well the communications gap was being bridged

    • Assessed progress toward achieving the intended con-tribution of “100% of referrals receive a feedback form”

    Workload data • Provided data to inform leaders of decisions regarding staffi ng requirements for the model

    • Provided a picture of when public health nurses were attending the maternity unit and for how long

    Document reviews • Provided information on how forms were actually being used by staff to inform re-design of the forms to be more useful

    • Assessed progress toward achieving the intended con-tribution of “100% of women who need complex care from at least 3 of the partners on the Transition Team”

    Observations at meetings

    • Allowed for the documentation of why decisions were made

    Interviews and focus groups with staff and leaders

    • Provided information about the outcomes experienced by the clients

    • Provided information on what was working and what challenges were being faced by staff to inform decisions about what practices to adopt and when alternative practices should be tested out

    • Assessed progress toward achieving the intended contribution of “100% of healthcare providers on the Perinatal Transition Teams report understanding of roles of all members of the team”

    Th e intention was to have the evaluator spend 0.4 full-time-equivalent days (15 hours/week) on the project. In reality there were times where the workload on her other projects had to be shift ed to accommodate the work required for this evaluation. Th ere were no additional resources provided for the prototype project, meaning all planning and evaluation work was conducted by reallocating existing staff . Some data collection and analysis was conducted by nursing staff working on the project (e.g., nurses recorded their time in a spreadsheet, a nurse analyzed some client data extracted from referral forms), but otherwise the evaluation work was conducted by the evaluator. Th us, there was not suffi cient evaluation capacity

  • Evaluating a Prototype Project 187

    CJPE 30.2, 177–194 © 2015doi: 10.3138/cjpe.209

    available to conduct some elements that would have enhanced the evaluation, such as interviewing individuals who did not work for the health authority (i.e., primary care providers and clients) about their experiences.

    DESCRIPTION OF CHALLENGES, HOW THEY IMPEDED THE EVALUATION PROCESS, AND HOW THE CHALLENGES WERE ADDRESSED

    What challenges did you face in conducting this evaluation? To what extent did you anticipate or could you have anticipated these challenges? As with any project occurring in a complex environment, numerous challenges were faced in this project. Th ese can be broken down into two types: challenges associated with the prototype project implementation and challenges associated with the DE approach. Th e challenges included (a) unfamiliarity and uncertainty associated with a prototype project, (b) bureaucratic hurdles in a large organiza-tion not aligning with the rapid pace of a prototype project, and (c) high turnover in leadership throughout the project. Th e nature of most of these challenges, with the exception of the high turnover in leadership, was anticipated; however, the extent to which they aff ected the work was underestimated.

    How did these challenges aff ect the implementation of the evaluation? How did you address each of these challenges?

    Unfamiliarity and uncertainty associated with a prototype project

    Most of those involved in this project had never been involved in a prototype project, and features of a prototype were not always clearly understood by the participants. Prototyping is a very diff erent way of working than most people in healthcare are used to. New healthcare programs are oft en introduced as pilot projects, where a fully formed protocol or procedure is tested in a setting, with little to no opportunity to focus on and capture learnings or to change or adapt the procedure. By contrast, it was diffi cult, especially in the beginning, for stake-holders to understand exactly what the Seamless Perinatal Healthcare Initiative was, as the model presented to staff at the start of prototype implementation was high-level, without specifi c details for procedures to be followed. In addition, staff were not provided with specifi c “rules” around what to do to operationalize the model, which made it diffi cult for the staff involved to understand if they were doing what they were “supposed” to be doing. In addition, concerns were raised about what, exactly, was being evaluated. Unfamiliarity with this way of working led to some distrust of the evaluator, who was viewed by some stakeholders as collecting information to report back to leadership for punitive reasons, instead of for the purpose of learning to improve the model being prototyped.

    To address these challenges, the evaluator was embedded in all levels of the governance structure including the IPT, where front-line staff and managers

  • 188 Snow, Urbina-Beggs, and Van Tent

    © 2015 CJPE 30.2, 177–194 doi: 10.3138/cjpe.209

    discussed their progress with the prototype (see Figure 2 for meeting frequency). Th ese meetings provided regular opportunities to use data from the prototype to inform decisions. In addition, she attended the two hospital sites to talk to staff , reiterate the prototyping purpose and process, listen to concerns, and answer questions. Th is allowed her to employ a variety of DE practices such as orienting staff to the prototyping process’s emergent nature and intervening through mod-erating discussions around their assumptions ( Table 5 ).

    Table 5. Developmental Evaluation

    Developmen-tal evaluation practice

    Description of the practice

    Example of how it was used in this project

    Orienting Helping stakehold-ers frame their work, including defi ning its ele-ments, testing their models, and under-standing their roles within a broader context

    Setting “intended contributions,” which were measurable progress markers that provided stakeholders with a goal to orient them in their work and a way to assess progress toward their goals. Moreover, an “intended contribu-tion” was set for each of the “3 Cs” of the vision (improving communication, coordination, and collaboration in the care of vulnerable pregnant women and new mothers) to frame the work around these important goals.

    Watching Attending to key moments, dynam-ics, power, and collective learning

    The evaluator noted high leadership turn-over, which coincided with a breakdown in the normal schedule of steering committee meetings, led to a gap in decision-making, and thus was stalling the project. To mitigate, there was a concerted eff ort to provide a formal orientation for the incoming project manager, including a transition period of working with the outgoing project manager and a tight collaboration with the evaluator.

    Sense-making Helping the group to analyze and give meaning to the data being col-lected

    The evaluator collected and analyzed data on the characteristics of the clients being re-ferred to the Transition Teams to gain a bet-ter understanding of the patient population being served. These data, which showed, for example, very diff erent referral patterns at the small community hospital compared to the large tertiary hospital, were brought to the implementation planning team and steering committee for discussion, inter-pretation, and determination of how these data would inform the future roll-out of the model at other hospitals.

  • Evaluating a Prototype Project 189

    CJPE 30.2, 177–194 © 2015doi: 10.3138/cjpe.209

    Developmen-tal evaluation practice

    Description of the practice

    Example of how it was used in this project

    Intervening Taking actions such as asking ques-tions, moderating discussion, re-minding, and sup-porting collabora-tion to infl uence the development of the program/intervention

    When a stakeholder expressed concern they would be punished for testing out a way of working that was diff erent from what they believed leadership wanted, the evaluator asked questions to reveal the assumptions being made. In this case, this stakeholder expressed the desire that the evaluator withhold from leadership the information that public health nurses were working directly with patients in the hospital for fear of being reprimanded by leadership. The evaluator led the group through a thought experiment by asking, “What would happen if you continued to work this way but did not tell the steering committee?” The group realized that because they had been seeing positive results, the steering committee would conclude that the positive results were coming from public health nurses not seeing patients in the hospital, and then they would fi nalize the model that way. The evaluator used this opportunity to remind the group that the purpose of the proto-type was to learn, and thus testing out a diff erent way of working was encouraged. This helped the stakeholder to see this work as providing evidence demonstrating this new way of working achieved the intended results and would inform leadership on decisions for the program. Thus, the evalua-tion supported stakeholders to analyze and synthesize learnings to make sense of them in the context of both the prototype sites and for future roll-out of the initiative to other sites.

    Bureaucratic hurdles in a large organization not aligning with the rapid pace of a prototype project

    In the prototype project, staff were making quick changes to how they did their work to test which ways of working were most eff ective. For example, staff tried out a process where the PHN working in the hospital did not see patients directly (working instead with inpatient maternity nurses on care planning) as

  • 190 Snow, Urbina-Beggs, and Van Tent

    © 2015 CJPE 30.2, 177–194 doi: 10.3138/cjpe.209

    opposed to working directly with the patients in the hospital (again, planning care as opposed to providing direct care). Similarly, as staff worked with the initial referral and feedback forms, they continually found ways to improve them, adapting the form by adding new fi elds to it as they worked. In contrast, systems within the health authority are much slower than a prototype (e.g., the turnaround time for a privacy impact assessment by the organization on changes made to the forms was 6–8 weeks versus staff making changes to the form in an ongoing manner).

    With respect to the evaluation process itself, data collection was challeng-ing, as the required data evolved along with other aspects of the prototype. For example, data required at the start (e.g., detailed client information to gain understanding of the patient population being referred to the transition team) was not necessarily required later on, so some data being collected was dropped once it was no longer necessary, and new data to collect (e.g., interviews/focus groups with staff about the ways in which they were working together) was added as needs arose. Moreover, attempting to collect data to assess the three “intended contributions” surfaced issues around the appropriateness of those measures for the prototype; as those issues were discussed with project staff , new “intended contributions” were established ( Table 3 ). Clear communication was necessary to ensure all stakeholders understood why data collection was changing during the evaluation.

    To address these challenges, it was necessary to balance the speed of changes inherent in a prototype project with not overwhelming and confusing participants by making too many changes at the same time. Moreover, since diff erent individuals assumed the role of PHN on the transition team on a given day, it was necessary to fi nd ways to clearly communicate changes in a timely manner to everyone involved. Having staff from both sites meet together as the IPT to discuss their process was helpful in this regard. It was important to be clear about what changes were being made and why, and for the evaluator to prioritize what data needed to be collected and in what timeframes, given the limited evaluation resources. In addition, a clinical nurse specialist stepped in to help with data collection and analysis to ease some of the workload for the lone evaluator. High turnover in leadership throughout the project

    An unexpected challenge was high leadership turnover, with six of the nine leaders engaged in the project, and the project manager, retiring or taking new positions during the 13 months of the prototype. Continuity and institutional memory was diffi cult to maintain with so many players and so much turnover. Some components identifi ed in early planning stages, on which future work was built, did not seem to have been eff ectively communicated to those who joined in the work later, resulting in some confusion. Moreover, there was no formal transition or orientation for new leaders taking over from previous individuals. In addition to leadership turnover, there was also signifi cant turnover of front-line staff engaged in the prototyping work, as new individuals were brought on

  • Evaluating a Prototype Project 191

    CJPE 30.2, 177–194 © 2015doi: 10.3138/cjpe.209

    to take part in the Seamless Perinatal Transition Teams as the work took shape. With respect to the evaluation itself, the high turnover in leadership presented a challenge both because it became unclear with whom to share evaluation fi ndings from the IPT and because new leaders needed to be introduced to the concepts of prototyping and DE in the midst of the project.

    Th rough the DE practice of watching, particularly around project structure ( Dozois et al., 2010 ; Langlois, Blanchet-Cohen, & Beer, 2013 ), the evaluator noted high leadership turnover, coinciding with a breakdown in the SC meeting sched-ule, leading to a gap in decision-making and stalling the project. To mitigate, there was a concerted eff ort to provide a formal orientation for the incoming project manager, including a transition period of working with the outgoing project manager and a tight collaboration with the evaluator.

    What should evaluators do to avoid these challenges to start off with? What would you recommend for others faced with similar challenges? As noted by Zimmerman et al. (2011) , “Evaluating complex interventions requires [. . .] a willingness to be uncertain at times and to know that being uncertain is crucial to the process” (p. viii). Prototype projects, by defi nition, involve more uncertainty and ambiguity at the start than traditional pilots. Some people are more comfortable with uncertainty than others, and staff should be supported specifi cally to work in this new way. Leadership should clearly, consistently, and continually communicate what a prototype is, and they should support staff through the uncertainty (e.g., focus on the concept that learning by doing is a key outcome of this process). An evaluator can assist in this process through facilitat-ing the development of a compelling vision that is motivating to stakeholders and that can help guide those working in situations of uncertainty ( Dozois et al., 2010 ). In addition, an evaluator could help reduce some uncertainty by facilitating the group to explicitly state some rules by which the front-line staff should operate during the prototyping process.

    When embarking on a prototype project, particularly with stakeholders unfamiliar with this style of working, it would be wise to provide signifi cant sup-port around change, uncertainty, and complexity. Th ough some work was done in this regard, it could have been improved by being more thoroughly planned upfront and provided in a timelier, ongoing manner, with particular care taken to orient new staff and leaders who become engaged throughout the project. In addition, more clearly defi ning the roles of those involved and ensuring there is adequate transition planning and orientation when turnover occurs would be benefi cial.

    Interestingly, when stakeholders were asked, “If you were to be involved in another such project from the beginning, what would you do diff erently?” the general consensus was, despite the challenges, prototyping was the ap-propriate method for the project; there was a sense the group needed to go through the testing and learning process to see how the work would unfold. It

  • 192 Snow, Urbina-Beggs, and Van Tent

    © 2015 CJPE 30.2, 177–194 doi: 10.3138/cjpe.209

    was noted that, in the past, projects were oft en created on paper and expected to be rolled out in a pilot-type fashion; however, once launched, many things were found not to work as planned, yet there was no process to capture learn-ings and make needed adaptations. It was felt with this initiative, despite it being diffi cult at times, the group needed to go through the process to arrive at where they needed to be. Many felt the right people were at the table—from the front-line nurse to upper management, and from the diff erent disciplines. Similarly, working together was seen as the most eff ective way to build trust and relationships.

    With a large, multisite prototype, it is impossible for a single evaluator to be present at diff erent sites as learning occurs. If additional evaluation resources are not available, training staff who are conducting the prototype on principles of DE could allow them to take a more active role in monitoring, evaluation, and using real-time fi ndings to inform the prototyping process.

    What, if any, are the systemic issues that the evaluation community should address? Demonstrating (or even defi ning) “success” is challenging with a prototype pro-ject. Decision makers oft en want to see quantitative outcome data, such as cost savings or improvements in health status, whereas prototyping is a developmental process in which learning is a primary intended outcome; the learnings are used to improve the model during the prototyping process. More research on ways in which DE ultimately contributes to outcomes in the long term could help support evaluators to demonstrate the value of this type of work and in advocating for this approach to evaluation for complex initiatives. Moreover, support in the devel-opment of clear and consistent messaging about the ways in which prototyping and DE are diff erent from traditional process or outcome evaluations, as well as mechanisms for building capacity in development evaluation, could be helpful.

    ACKNOWLEDGEMENTS Th e authors would like to express their gratitude to the many individuals who contributed to the success of this project, including the members of the working group, advisory and steering committees, and the staff and management from the two prototype sites. Special acknowledgement goes to retired Public Health Director, Linda Bachmann, whose vision for a seamless, patient-focused client journey was the foundation of this project.

    REFERENCES American Society for Quality . ( n.d. ). Continuous improvement . Retrieved from http://asq.

    org/learn-about-quality/continuous-improvement/overview/overview.html Begun , J. W. , Zimmerman , B. , & Dooley , K. ( 2003 ). Health care organizations as complex

    adaptive systems . In S. M. Mick & M. Wyttenbach (Eds.), Advances in health care

    http://asq.org/learn-about-quality/continuous-improvement/overview/overview.htmlhttp://asq.org/learn-about-quality/continuous-improvement/overview/overview.html

  • Evaluating a Prototype Project 193

    CJPE 30.2, 177–194 © 2015doi: 10.3138/cjpe.209

    organization theory (pp. 253 – 288 ). San Francisco, CA : Jossey-Bass . Retrieved from http://change-ability.ca/fi les/Complex_Adaptive.pdf

    Chopova , D. ( 2011 ). Literature review of best practices in perinatal transitions of care . Surrey, BC : Fraser Health .

    Dozois , E. , Langlois , M. , & Blanchet-Cohen , N. ( 2010 ). DE 201: A practitioner’s guide to developmental evaluation. Montreal: J. W. McConnell Family Foundation and the In-ternational Institute for Child Rights and Development. Retrieved from http://www.mcconnellfoundation.ca/en/resources/publication/de-201-a-practitioner-s-guide-to-developmental-evaluation

    Ferguson , C. ( n.d. ). Applying design thinking to Ontario healthcare . Retrieved from http://bridgeable.com/applying-design-thinking-to-ontario-healthcare/

    Fraser Health . ( 2011 ). Gap analysis: Fraser Health perinatal transitions—current practice/needs assessment . Surrey, BC : Author .

    Langlois , M. , Blanchet-Cohen , N. , & Beer , T. ( 2013 ). Th e art of the nudge: Five prac-tices for developmental evaluators. Canadian Journal of Program Evaluation , 27 ( 2 ), 39 – 59 .

    Patton , M. Q. ( 2011 ). Developmental evaluation: Applying complexity concepts to enhance innovation and use . New York, NY : Guilford Press .

    Rey , L. , Tremblay , M. C. , & Brousselle , A. ( 2014 ). Managing tensions between evalu-ation and research: Illustrative cases of developmental evaluation in the con-text of research. American Journal of Evaluation , 35 ( 1 ), 45 – 60 . http://dx.doi.org/10.1177/1098214013503698

    Zimmerman , B. , Dubois , N. , Houle , J. , Lloyd , S. , Mercier , C. , Brouselle , A. , & Rey , L. ( 2011 ). How does complexity impact evaluation? Canadian Journal of Program Evaluation , 26 ( 3 ), v – xx .

    AUTHOR INFORMATION M. Elizabeth Snow , PhD, MBA, CE, was the Evaluation Specialist in Public Health at the Fraser Health Authority at the time of this project. She holds a BSc(Hons) from McMaster University, an MSc from the University of Guelph, a PhD from the University of British Columbia (UBC), an MBA from the Sauder School of Business at UBC, and a Credentialed Evaluator designation from the Canadian Evaluation Society. She is currently the Program Evaluation Lead for the Centre for Health Evaluation and Outcomes Sciences (CHÉOS) and the Clinical and Systems Transformation Project in Vancouver, BC. As well, she is an instructor at the Justice Institute of British Columbia and an adjunct professor in the Fac-ulty of Health Sciences at Simon Fraser University. Her work focuses on bridging the gap between research and evaluation evidence and health services delivery. Michelle Urbina-Beggs , RN, BScN, MN, has been working in the fi eld of Population and Public Health since 2002 and is currently the Clinical Nurse Specialist for Population and Public Health for the Fraser Health Authority in British Columbia, Canada. She holds a BScN from the University of Victoria and an MN from the University of Southern Queens-land, Australia. Michelle’s role supports the establishment and advancement of perinatal, child, and family health public health services and is focused on translating theory into evidence-informed practice.

    http://change-ability.ca/fi les/Complex_Adaptive.pdfhttp://www.mcconnellfoundation.ca/en/resources/publication/de-201-a-practitioner-s-guide-to-developmental-evaluationhttp://bridgeable.com/applying-design-thinking-to-ontario-healthcare/http://bridgeable.com/applying-design-thinking-to-ontario-healthcare/http://dx.doi.org/10.1177/1098214013503698http://www.mcconnellfoundation.ca/en/resources/publication/de-201-a-practitioner-s-guide-to-developmental-evaluationhttp://www.mcconnellfoundation.ca/en/resources/publication/de-201-a-practitioner-s-guide-to-developmental-evaluationhttp://dx.doi.org/10.1177/1098214013503698

  • 194 Snow, Urbina-Beggs, and Van Tent

    © 2015 CJPE 30.2, 177–194 doi: 10.3138/cjpe.209

    Tamara Van Tent , RN, MHS, has been involved in maternal newborn service and program delivery in various capacities for 30 years. She is currently Director of Clinical Operations for the Maternal Infant Child and Youth Program in the Fraser Health Authority, British Columbia, Canada. Tamara’s role includes leadership for operations, quality, and strategic planning for maternal, newborn, and pediatric services across four acute care hospitals. Tamara is the Acute Clinical Program Director sponsor of the Seamless Perinatal Project discussed in this article.

    Challenges in Evaluating a Prototype Project in a Large Health Authority: Lessons LearnedDESCRIPTION OF CASE AND EVALUATION CONTEXTWhy Was the Evaluation Conducted? What Did the Client Want to Learn?What resources (time, money, in-kind, etc.) were available for conducting the evaluation? Were they suitable for answering the evaluation questions?

    DESCRIPTION OF CHALLENGES, HOW THEY IMPEDED THE EVALUATION PROCESS, AND HOW THE CHALLENGES WERE ADDRESSEDWhat challenges did you face in conducting this evaluation? To what extent did you anticipate or could you have anticipated these challenges?How did these challenges affect the implementation of the evaluation? How did you address each of these challenges?Unfamiliarity and uncertainty associated with a prototype projectBureaucratic hurdles in a large organization not aligning with the rapid pace of a prototype projectHigh turnover in leadership throughout the project

    What should evaluators do to avoid these challenges to start off with? What would you recommend for others faced with similar challenges?What, if any, are the systemic issues that the evaluation community should address?

    ACKNOWLEDGEMENTSREFERENCESAUTHOR INFORMATION

    /ColorImageDict > /JPEG2000ColorACSImageDict > /JPEG2000ColorImageDict > /AntiAliasGrayImages false /CropGrayImages false /GrayImageMinResolution 200 /GrayImageMinResolutionPolicy /OK /DownsampleGrayImages true /GrayImageDownsampleType /Bicubic /GrayImageResolution 450 /GrayImageDepth -1 /GrayImageMinDownsampleDepth 2 /GrayImageDownsampleThreshold 1.50000 /EncodeGrayImages true /GrayImageFilter /DCTEncode /AutoFilterGrayImages true /GrayImageAutoFilterStrategy /JPEG /GrayACSImageDict > /GrayImageDict > /JPEG2000GrayACSImageDict > /JPEG2000GrayImageDict > /AntiAliasMonoImages false /CropMonoImages false /MonoImageMinResolution 595 /MonoImageMinResolutionPolicy /OK /DownsampleMonoImages true /MonoImageDownsampleType /Bicubic /MonoImageResolution 1200 /MonoImageDepth -1 /MonoImageDownsampleThreshold 1.50000 /EncodeMonoImages true /MonoImageFilter /CCITTFaxEncode /MonoImageDict > /AllowPSXObjects true /CheckCompliance [ /None ] /PDFX1aCheck false /PDFX3Check false /PDFXCompliantPDFOnly false /PDFXNoTrimBoxError true /PDFXTrimBoxToMediaBoxOffset [ 0.00000 0.00000 0.00000 0.00000 ] /PDFXSetBleedBoxToMediaBox true /PDFXBleedBoxToTrimBoxOffset [ 0.00000 0.00000 0.00000 0.00000 ] /PDFXOutputIntentProfile (U.S. Web Coated \050SWOP\051 v2) /PDFXOutputConditionIdentifier (CGATS TR 001) /PDFXOutputCondition () /PDFXRegistryName (http://www.color.org) /PDFXTrapped /False

    /Description > /Namespace [ (Adobe) (Common) (1.0) ] /OtherNamespaces [ > > /FormElements true /GenerateStructure false /IncludeBookmarks false /IncludeHyperlinks false /IncludeInteractive false /IncludeLayers false /IncludeProfiles false /MarksOffset 6 /MarksWeight 0.250000 /MultimediaHandling /UseObjectSettings /Namespace [ (Adobe) (CreativeSuite) (2.0) ] /PDFXOutputIntentProfileSelector /UseName /PageMarksFile /RomanDefault /PreserveEditing true /UntaggedCMYKHandling /LeaveUntagged /UntaggedRGBHandling /LeaveUntagged /UseDocumentBleed true >> ] /SyntheticBoldness 1.000000>> setdistillerparams> setpagedevice


Recommended