+ All Categories
Home > Documents > Session Outcomes

Session Outcomes

Date post: 23-Jan-2016
Category:
Upload: kenyon
View: 42 times
Download: 0 times
Share this document with a friend
Description:
Technical Assessment Development and Validation: Methods for Ensuring The Utility, Validity and Reliability of Technical Skill Assessment Systems. Session Outcomes. Build understanding of key criteria for technical skill assessments: Utility Validity Reliability - PowerPoint PPT Presentation
Popular Tags:
32
Technical Assessment Development and Validation: Methods for Ensuring The Utility, Validity and Reliability of Technical Skill Assessment Systems
Transcript
Page 1: Session Outcomes

Technical Assessment Development and Validation: Methods for Ensuring The Utility, Validity and Reliability of Technical Skill Assessment Systems

Page 2: Session Outcomes

Session OutcomesSession Outcomes

• Build understanding of key criteria for technical skill assessments:

Utility Validity Reliability

• Understand a state-led process for developing a technical assessment system that meets such criteria.

Page 3: Session Outcomes

Carl D. Perkins Career and Carl D. Perkins Career and Technical Education Act of 2006Technical Education Act of 2006

• Each state established a performance accountability system with multiple measures of student learning, program completion, and transitions to further education, employment and the military

• Perkins III allowed wide flexibility in how to measure “technical skill attainment”

• Perkins IV requires a more focused assessment approach for technical skill attainment

Page 4: Session Outcomes

Technical Skill Attainment -- Technical Skill Attainment -- SecondarySecondary

• Sec 113 (b)(2)(A)

• …”core indicators of performance…that are valid and reliable… measures of each of the following:”

• “Student attainment of career and technical skill proficiencies, including student achievement on technical assessments, that are aligned with industry-recognized standards, if available and appropriate.”

Page 5: Session Outcomes

Technical Skill Attainment -- PostsecondaryTechnical Skill Attainment -- Postsecondary

Sec 113 (b)(2)(B)…”core indicators of performance…that are valid and

reliable…measures of each of the following:”“Student attainment of career and technical skill

proficiencies, including student achievement on technical assessments, that are aligned with industry-recognized standards, if available and appropriate.”

“Student attainment of an industry-recognized credential, a certificate, or a degree.”

Page 6: Session Outcomes

Critical Features …

a) Measure what is important

b) Useful and timely feedback to stakeholders

c) Fair, consistent and accurate measures (e.g., reliable & valid assessments)

Key Concepts

Utility

Validity

Reliability

Page 7: Session Outcomes

Utility: Something useful or designed for use

Some Core Assumptions1. We have to do this, so let’s do it in a way that is

going to be maximally useful to our stakeholders.

2. Assessment systems should ultimately influence and reflect what is occurring in the educational setting(s).

3. Without buy-in, items 1 & 2 will never happen. 4. A systematic process for stakeholder

involvement and communication must be explicitly planned and built-in.

Page 8: Session Outcomes

Validity: To what extent does the assessment measure what it is supposed to measure?

Commonly used methods… Face and content validity Construct validity (convergent, divergent, factor

analytic techniques, etc.)

Criterion-related validity (concurrent validity, predictive validity)

Page 9: Session Outcomes

Reliability refers to the stability or consistency of assessment results. Does the assessment measure yield consistent results across different raters, different periods of time, different samples of tasks, and so forth.

Commonly used methods Internal consistency reliability Test-Retest reliability Inter-rater reliability Others: (equivalency/parallel content, expert- rater reliability, etc.)

Page 10: Session Outcomes

Wyoming CTE Assessment Wyoming CTE Assessment Project GoalsProject Goals

• Establish shared expectations as to what students should know and be able to do in Wyoming’s CTE programs

• Develop a valid and reliable CTE assessment system

• Ensure the system provides useful, timely and accurate feedback to teachers, administrators, students and employers

Page 11: Session Outcomes

Options to fulfill Perkins IV are:Options to fulfill Perkins IV are:

Use Industry-Based Certifications or other Use Industry-Based Certifications or other standardized-assessments –standardized-assessments –

AND/ORAND/OR

Develop valid and reliable assessments Develop valid and reliable assessments through a statewide collaborative processthrough a statewide collaborative process

Page 12: Session Outcomes

Challenges and ConsiderationsChallenges and Considerations

• Access to assessment data to improve classroom instructionAccess to assessment data to improve classroom instruction• Dealing with the expense of buying IBC’sDealing with the expense of buying IBC’s• Making sure the IBC’s match up to the course contentMaking sure the IBC’s match up to the course content• Making sure the IBC is valuable to employers and the job Making sure the IBC is valuable to employers and the job

marketmarket• Getting data from externally administered examsGetting data from externally administered exams• Deciding when to assess (end-of-program or course by Deciding when to assess (end-of-program or course by

course)course)• Assessments that are appropriate to various program Assessments that are appropriate to various program

structures and goalsstructures and goals• Assessing CTE skills AND employability skillsAssessing CTE skills AND employability skills

Page 13: Session Outcomes

Putting First Things FirstPutting First Things First

FIRST, decide WHAT to assess

THEN decide

HOW to assess

Page 14: Session Outcomes

Source of StandardsSource of Standards

State standardsSCANS

State standardsSCCI K&S

State standardsSCCI K&S

State standards

IndustryState standardsdefacto – texts

and tests

Courtesy of Steve Klein & MPR Associates

Page 15: Session Outcomes

Can One Assessment Measure Can One Assessment Measure It All?It All?

No matter the approach, it is inherent that the program will include applied academic skills, employability skills, cluster- and pathway-level skills, and program/occupation skills. When considering assessments, consider if one assessment can adequately measure all those skills.

Industry certification

test

State test from National Item

Banks

Commercial employability

skills test

Program of Study

Page 16: Session Outcomes

Setting up the StructureSetting up the Structure

Assessment Project Advisory Group• 20-25 participants, CTE administrators, community college

administrators, teachers from various clusters, state agency staff,

• Provide general input on development process • Liaison to the education communities at secondary and

postsecondary levels• Identify and prioritize clusters for development in remainder

of project• Meet in-person and through webinars, 2-3 times per year

Page 17: Session Outcomes

Setting up the StructureSetting up the Structure

Business/Industry Advisory Group• Cross-section of business/industry representatives.• Should include representatives from each of the three initial clusters

(Agriculture, Construction, Manufacturing)• Provide general input on development process from

business/industry perspective• Liaison to the business communities across the state• Review, provide input, affirm content developed by Cluster/Pathway

Work Team• Advise on raising value of CTE and the CTE assessment system

within Wyoming business/industry.• Meet in-person and through webinars, 2-3 times per year

Page 18: Session Outcomes

Setting up the StructureSetting up the StructureCluster/Pathway Work Groups• 7-10 content experts in each of three clusters: Agriculture. Construction,

Manufacturing• Provide input on the priority competencies to include in the assessment

system• Assist in identifying the relative usefulness and applicability of existing

assessments• Provide input on any state-developed assessments that are determined to

be necessary• Kick-off briefings on March 7, 2008.• Work sessions March 16-20, 2008.• In-person and webinar follow-up sessions through June 2008.• Optional involvement in assessment pilot phase.

Page 19: Session Outcomes

Identifying Competencies and ObjectivesIdentifying Competencies and Objectives

March 08 • Convene initial Cluster/Pathway Working Groups (CPWG).• Each CPWG identifies core competencies (technical,

academic, employability) that need to be assessed in each Cluster/Pathway.

April-May 08• Draft Competencies are completed by CPWG and posted

online for review.• Other WY teachers and faculty invited to review and

comment on Draft Competencies.• Cluster/Pathway Competencies finalized

Page 20: Session Outcomes

Identifying Test Items and Identifying Test Items and Assessment OptionsAssessment Options

May-June 08• 5/12/08, Manufacturing and Arch/Construction CPWG’s meet

to review sample test items for the Competencies

June 08 • Agriculture/Natural Resources CPWG meets to review

sample test items for competencies• Consultant team gathers information on assessment

resources (NOCTI, SkillsUSA, industry groups) and delivery system options

• Consultant team completes feasibility report for assessment development phase.

Page 21: Session Outcomes

Pilot Testing Assessments and Pilot Testing Assessments and Next StepsNext Steps

Fall 2008 – Spring 2009• Development and pilot testing of first phase assessments for

initial clusters• Possibly begin to work with additional Cluster/Pathway Work

Groups to identify Essential Core competencies for other areas

Page 22: Session Outcomes

Example – the Utah CTE assessment Example – the Utah CTE assessment systemsystem

Page 23: Session Outcomes

Overview of Assessment Development Process

• Identify competencies and objectives. • Decide what to assess and how (e.g., develop an

assessment blueprint). • Feasibility phases examine existing options• Pilot assessments and conduct necessary

analyses to document technical quality of assessments.

• Finalize assessments, delivery, key features the system must possess.

Page 24: Session Outcomes

What is an Assessment Blueprint?

An assessment blueprint helps us determine what should be covered in the assessment(s) as well as the number of test items that should be included in each category. You can also use it to help determine the total length of the test as well as the types of items to be included.

Page 25: Session Outcomes

Example: NOCTI Experienced Worker Assessment Blueprint

Areas covered in the Building Construction Occupations, Written Assessment:

29% Carpentry17% Electrical7% Plumbing4% Math6% Metal Work/Guttering10% Painting & Decorating8% Building Code & Safety

19% Masonry

Page 26: Session Outcomes

Example: Texas Education AgencyTAKS 9th Grade Reading

Objective 1: Reading and basic understanding

9 multiple-choice items

Objective 2: Reading – literacy elements and techniques

12 multiple-choice items1 short answer item

Objective 3: Reading – analysis and critical evaluation

12 multiple-choice items2 short answer items

TOTAL NUMBER OF ITEMS

33 multiple-choice items3 short answer items

Page 27: Session Outcomes

One last example …

A sample assessment blueprint from a test on human

geography is provided below.:  

The construction of the blueprint is a useful process as it helps to ensure that items are produced that cover both the content of the program and the educational objectives.  It can also allow the balance of 'worth' of individual items to be determined. 

Area Knows common terms

Knows specific facts

Understand principles

Applies principles

Interprets charts and graphs

Total

Food 1 2 1 2 2 8

Clothing 2 1 0 1 1 5

Shelter 1 0 1 2 1 5

Page 28: Session Outcomes

Developing an Assessment Blueprint

• What is the relative emphasis you want to place on cluster level competencies versus pathway level competencies?

• What is the relative emphasis you want to place on areas within the pathway?

• Within a cluster/pathway, are there objectives that are a greater priority for you to measure than others?

• What parameters do you wish to set for the total length and duration of the test?

• What are your thoughts regarding the distribution and use of different types of assessment items across the areas? How many multiple choice, short-answer/constructed response, or performance tasks? (Note: Blooms Taxonomy, etc.)

Page 29: Session Outcomes

Some factors to consider when examining potential existing assessments

• Alignment: Do the items align or match up with the competencies/objectives we’ve identified as important? (e.g., Is the assessment measuring what we want it to measure?

• Ease of use: Is it manageable for teachers and students (e.g., administration method, directions clear, requirements in terms of resources, etc.)

• Administration Time • Cost • Flexibility• Fairness• Content in terms of assessment items (see next page)

Page 30: Session Outcomes

Some characteristics of good assessment items …

• Each item has a specific purpose and is designed to test a significant learning outcome.

• Items are clear (avoid irrelevant material, also language should be clear and easy to understand or else you may be measuring student comprehension of English rather than the trait you wish to measure).

• Contains plausible distractors• Do the items contained within the assessment employ

multiple methods to provide a more complete picture of student knowledge and/or skills?

• Do questions discriminate between the more able and less able students? Do they allow students to go well beyond the threshold requirements if they are able to?

Page 31: Session Outcomes

Key features of the assessment system?

• Timing of assessments• Types of scores produced• Reporting (ongoing documentation? Access?)• Order of presentation (items presented in a specific order,

randomly, etc.)• If online, desired features? (security/access, timed access,

etc.)• Other things you want to be sure the system has or is able to

do???

As a teacher, what are the key features this CTE assessment system should have in order to make it really useful for you and your students?

Page 32: Session Outcomes

For more information, contact:For more information, contact:

• Mariam Azin: Mariam Azin: [email protected]@presassociates.com

www.presassociates.comwww.presassociates.com

• Hans Meeder:Hans Meeder: [email protected] [email protected]

www.MeederConsulting.comwww.MeederConsulting.com


Recommended