+ All Categories
Home > Documents > Introducing the Computer-Based Test for the National ...

Introducing the Computer-Based Test for the National ...

Date post: 21-Mar-2022
Category:
Upload: others
View: 2 times
Download: 0 times
Share this document with a friend
22
1. Research Background 3. Research Content 2. Research Goal 4. Policy Recommendations Research Report 2020 Introducing the Computer-Based Test for the National Assessment of Educational Achievement: A Developmental Study for the e NAEA Head Researcher│Jaebong Lee
Transcript

1. Research Background

3. Research Content

2. Research Goal

4. Policy Recommendations

Research Report 2020

Introducing theComputer-Based Test for the

National Assessmentof Educational Achievement:

A Developmental Studyfor the eNAEA

Head Researcher│Jaebong Lee

Research Report 2020

PIM 2021-8-3

III. Introducing the Computer-Based Test

for the National Assessment of Educational

Achievement: A Developmental Study for

the eNAEA

KICE Research Report 2020

Published Date August 27, 2021

Publisher Tae-Joong Gahng

Publishing AgencyKorea Institute for Curriculum and Evaluation

Address8, Gyohak-ro, Deoksan-eup, Jincheon-gun, Chungcheongbuk-do, Republic of Korea

Telephone +82-43-931-0114

Fax +82-43-931-0884

Home Page http://www.kice.re.kr

Printing Company KS SENSATION (+82-2-761-0031)

ISBN 979-11-91582-35-2 95370 979-11-91582-44-4 95370 (Set)

PIM 2021-8-3

※ Unauthorized copy of contents of this report is prohibited.

ISBN 979-11-91582-35-2 95370

ISBN 979-11-91582-44-4 95370 (Set)

KICE Research Report 2020

IIIIntroducing the

Computer-Based Test for the National Assessment of

Educational Achievement:

A Developmental Study for the eNAEA

Head Researcher | Jaebong Lee

Co-researchers | Junshik Kim, Jiseon Park, Kyunghee Sung,

Kwangsang Lee, Sora Lee, Hyeyun Jung, Soyoung Choi,

Kamyoung Kim (Kyungpook National University),

Yumin Ahn (Kongju National University),

Minsu Ha (Kangwon National University)

KICE Research Report 2020

4

1. Research Background

◈ Need to establish methods of assessment in line with the achievement standards

and competency-based assessment introduced in the 2015 Revised National

Curriculum and the international trend toward the introduction of computer-based

assessment

◈ Need to search for a new direction of the National Assessment of Educational

Achievement (NAEA) that can accommodate the educational innovations for

the future society

○ With the dawn of the Fourth Industrial Revolution and the transition toward

an intelligent information society, countries around the world are stepping

up their efforts to nurture and assess new competencies for the present

demands. Likewise, Korea has established the importance of competency in

the subject curriculum and general guidelines for the 2015 Revised National

Curriculum.

○ Large-scale and international assessment of educational achievement is

generally working toward computer-based assessment, both to provide the

effective evaluation of the competencies required for the future society and

as a method of overcoming the limitations of the traditional pencil-and-paper

assessment.

○ Likewise, NAEA must provide a holistic evaluation of its contents to guarantee

the nurturing of creative talent demanded by the society of the future and

look for new methods of assessment items in terms of methodology to guarantee

the appropriate measurement of student competency.

◈ Need to establish specific methods of computer-based assessment development

in order to support the policy of the Ministry of Eduction to fully implement the

electronic National Assessment of Educational Achievement (eNAEA) in 2022.

○ In March 2019, the Ministry of Education of Korea presented a plan for the

full administration of computer-based assessment in 2022 as a plan to

Introducing the Computer-Based Test for the National Assessment of Educational Achievement:A Developmental Study for the eNAEA

5

improve NAEA (Ministry of Education, 201p9).

○ Providing an effective response to this policy requires the establishment of

methods for implementing transitory strategies for computer-based

assessment. Concrete methods must also be established for the implementation

of the eNAEA, including new assessment and subject-specific items.

2. Research Goal

○ This study seeks to establish the methods of introducing the computer-based

test for the NAEA, which are in line with the recent technological

development and demands from social and political sectors, including the

entrance into an intelligent information society and the recent revision of

the National Curriculum.

○ This study proposes an effective test development procedure to prepare for

the upcoming 2022 eNAEA.

○ By exploring the strategies to facilitate a stable transition to a computer-

based assessment scheme, this study establishes the directions for eNAEA

test item development, devises valid item types, develops subject-specific

plans and sample items, and establishes methods of developing test items in

line with the requirements of computer-based assessment.

3. Research Content

◈ Literature review and direction-setting for the transition toward computer-based

assessment

○ Implications from the analysis of transition underwent through PISA, TIMSS,

NAEP, and NAPLAN

― These cases involved a gradual transition toward computer-based assessment,

and by doing so, realized the competency-based assessment, hitherto thought

to be difficult in pencil-and-paper assessment, by introducing the innovative

KICE Research Report 2020

6

test items that allow test-takers to interact with the computer.

― In the case of NAEA, where the transition must be completed in a relatively

short period, there is a need to exercise the utmost care in preparing the

transition from the current paper-based test and development of new

innovative items and test-set designs to minimize the trial-and-error

process during the transition.

○ Implications from the literature review of the transition to computer-based

assessment

― Studies that discussed the opportunities and risks involved with the

transition of large-scale assessment to computer-based formats (Bridgeman,

2009; Bryant, 2017; de Klerk, 2012; Drasgow & Mattern, 2006; Reich &

Petter, 2009; Sireci & Zenisky, 2006; Strain-Seymour et al., 2009; van Lent,

2009) were analyzed in this study.

― Opportunities associated with the transition toward computer-based

assessment are improvement in the assessment through the introduction of

innovative items, improvement in student motivation and convenience, and

improvement of test administration efficiency through the electronification

of the system. Risks include issues in problems of proving the enhancement

of the construct validity, possibility of decreasing measurement efficiency,

as well as increase of test administration costs and security-related

management points.

― Transition strategies that are deemed suitable for the transition of NAEA

to a computer-based system: 1) strategies involving the innovative change

of the items themselves beyond the mere improvement of the system

efficiency, which can help the assessment system reduce the gap between

the assessment and real world, and induce educational innovations; and 2)

strategies that maximize the connection between the conventional

pencil-and-paper assessment and computer-based assessment and guarantee

its stable implementation.

Introducing the Computer-Based Test for the National Assessment of Educational Achievement:A Developmental Study for the eNAEA

7

○ Setting the direction for test development in eNAEA

― [Figure I-1] describes the transition trajectory from a conventional pencil-

and-paper-based system to the recent introduction of competency- and

computer-based assessment.

Ability-Based Assessment

Paper-and-Pencil Assessment

Traditional Knowledge-Based Assessment

Computer-Based Assessment

Direction of NAEA development

The 2009 Revised National Curriculum based paper-and-pencil assessment

The 2015 Revised National Curriculum based paper-and-pencil assessment

Research area for the present study

Method

Content

[Figure I-1] Direction of developments in NAEA following the introduction of

competency- and computer-based assessment

― Based on the analysis of international and overseas cases of the transition

to computer-based assessment, opportunities and risks associated with the

transition, and transition strategies, this study established eight aspects in

the direction-setting of the eNAEA test item development.

― Directions for eNAEA test item development: ① Develop items that are

appropriate to the goal of NAEA (examining the quality of the curriculum).

② Develop items that clearly represent the measurement construct. ③

Apply technical features only when the purpose of application is clear. ④

Develop items that encourage interactivity. ⑤ Apply technical features in

a step-wise fashion to accommodate students’ technical ability. ⑥ Develop

subject-specific items that are typical and applicable to the school

KICE Research Report 2020

8

context. ⑦ Develop items that maximize the utility of automatic scoring.

⑧ Develop items that secures stable test administration.

― These eight directions serve as basic principles of item format development

or system design and can be used as the criteria in the important

decision-making processes.

◈ Investigating and setting eNAEA item types

○ Types of test items utilizing technical features were set, in line with the

characteristics with NAEA under the 2015 Revised National Curriculum that

emphasizes competency.

― A list of major technical features applied in eNAEA can be found in <Table

I-1> and <Table I-2>.

<Table I-1> Technical features for response actions

Technical Feature Example a DescriptionSubject-Specific Utility

K SS M S E

Radio Button∙ Multiple choice question with one answer out

of five possible choices▲ ▲ ▲ ▲ ▲

Check Box∙ Multiple choice question with two or more

answers out of multiple (three to five) options▲ ▲ ▲ ▲ ▲

Matching∙ Matching the items in line with content

relations, information or other criteria▲ ▲ ▲ ▲ ▲

Pull-Down

∙ Inserting information by choosing one of the

answers from the pull-down menu

∙ Other names: Inline choice, drop-down

menu

▲ ▲ ▲ ▲ ▲

Hot Spot

∙ Clicking a specific area in a figure or text

(number, word, phrase, sentence, etc.)

∙ Other names: Figural response for nontextual

responses

▲ ▲ ▲ ▲ ▲

Drag and Drop

∙ Choosing an image or a word by clicking and

dragging to the answer field

∙ Includes the task of re-ordering the series of

numbers, words, phrases, sentences, etc.

▲ ▲ ▲ ▲ ▲

Introducing the Computer-Based Test for the National Assessment of Educational Achievement:A Developmental Study for the eNAEA

9

<Table I-2> Technical features for interactional activities

Technical Feature Example a DescriptionSubject-Specific Utility

K SS M S E

Graphing and

Drawing

∙ Includes connecting two points on a graph

with a line

∙ Includes free drawing functions on a separate

field

▽ ▲ ▲ ▲ ▽

Short Answer∙ Inputting characters or numbers into the text

box using the keyboard▲ ▲ ▲ ▲ ▲

Writing

Response

∙ Answers that are longer than a short answer

∙ Provided with a separate editing function

(cutting, pasting, underline, italics, etc.) for

free text response

▲ ▲ ▲ ▲ ▲

Math Symbols

∙ Mostly used in the math tests

∙ Can be used to enter chemical reaction

equations in the science tests

∙ Requires a separate math symbol input board

▽ ▽ ▲ ▲ ▽

Table

Completion

∙ Inputting numbers, symbols, characters, and

other answers into the answer field in the

table using a keyboard

▽ ▲ ▽ ▲ ▽

a: Examples provided include modified versions of the relevant examples from previous tests.

▲: Very likely to be used. ▽: Less likely to be used

K: Korean language, SS: Social Studies, M: Mathematics, S: Sciences, E: English

Technical

Feature

Level of

InteractivityDescription

Subject-Specific Utility

K SS M S E

Simulation High

∙ Identifying the relationship between variables by

manipulating tools or variables or predicting the

result by examining the function and significance

of the tool

▽ ▲ ▲ ▲ ▽

Tool

Manipulation

Moderate

– High

∙ Measuring an object presented in the test material

by manipulating a tool (protractor, ruler, calculator,

slider, etc.)

▽ ▽ ▲ ▲ ▽

Information

SearchHigh

∙ Searching for information through dictionaries or

the internet▲ ▲ ▽ ▽ ▲

Virtual

DialogueHigh

∙ Communicating with the computer based on a

dialogue scenario algorithm▲ ▽ ▽ ▽ ▲

Hypertext/

HyperlinkModerate

∙ Interacting with nonlinear, noncontinuous information

made up of links between documents, words, and

images

▲ ▲ ▲ ▲ ▲

Tab Low∙ Interacting with various information in different

formats or of different characteristics ▲ ▲ ▲ ▲ ▲

KICE Research Report 2020

10

○ Main item types for eNAEA were selected in the two aspects of test-taker

response actions and interactional activities, with consideration of the

efficiency and validity of the assessment tool.

― Considering that computer-based assessment systems allow for the utilization

of different technical features in student interactional activities and answer

areas (refer to [Figure I-2]), item types were set according to the two aspects

of test-taker response (refer to <Table I-3>) and interactional activities (refer

to [Table I-4]).

Instructions

Domain of the Interactional Activities

Domain of the Response Actions

[Figure I-2] Example of eNAEA item structure

Technical

Feature

Level of

InteractivityDescription

Subject-Specific Utility

K SS M S E

Pop-Up Low

∙ Interacting with information by placing the cursor

over events, which generate temporary windows

that disappear after the pointer is moved

▲ ▲ ▽ ▽ ▲

Multimedia Low∙ Presenting or searching data in the form of

animation, audio, or video▲ ▲ ▲ ▲ ▲

▲: Very likely to be used. ▽: Less likely to be used.

K: Korean language, SS: Social Studies, M: Mathematics, S: Sciences, E: English

Introducing the Computer-Based Test for the National Assessment of Educational Achievement:A Developmental Study for the eNAEA

11

<Table I-3> eNAEA item types in the aspect of response actions

Category eNAEA Item Type Technical Feature Notes

Selected

Response

1Multiple

Choice

Radio Button,

Check Box

∙ Includes the five-response multiple choice

items from the conventional pencil-and-paper

assessment

∙ Includes other multiple choice items

2Expanded

Choice

Hot Spot,

Pull-Down

∙ Can largely replace conventional multiple choice

item

∙ Removes the unnecessary textualization process

in the formulation of response

∙ Includes a pull-down feature, where the list of

choices is provided in the material itself

∙ Includes multiple choices in the “hot spot”

function

3Connecting

Tasks

Matching,

Drag and Drop

∙ Can largely replace conventional multiple choice

items

∙ Removes unnecessary textualization process in

the formulation of response

∙ Includes items involving classification according

to specific criteria

∙ Includes multiple choices in the matching

function

4 Ordering Drag and Drop

∙ Organizing the order of number, words, phrases,

sentences, or images, via clicking (selecting) or

using the drag and drop function

Constructed

Response

5 Short Answer

Table Completion,

Short Answer

Input, Math

Symbols Input

∙ When simple keyboard input is use, identical to

the conventional constructed response items

from the pencil-and-paper assessment

6Extended

Writing

Extended Writing

Input, Math

Symbols Input

∙ When simple keyboard input is use, identical to

the conventional constructed response items

from the pencil-and-paper assessment

∙ Includes writing using document editing functions

7Correction

Tasks

Hot Spot + Short

Answer/Extended

Writing/Math

Symbols Input

∙ Choosing and correcting incorrect contents

8Completing a

Graph/Figure

Graphing and

Drawing,

Drag and Drop

∙ Includes drawing or moving graphs

∙ Includes free drawing

∙ Includes the completion of incomplete items

(e.g. chemical structure models of sciences test

KICE Research Report 2020

12

<Table I-4> eNAEA item types in the aspect of interactional activities

eNAEA Item Type Technical Feature Notes

1Simple

Presentation

∙ Similar to the conventional pencil-and-paper item, static

screen, 2D plane

2Utilizing

Information

Information Pop-Up,

Information Search,

Tab,

Hypertext/Hyperlink

∙ Overcomes limitations in space allocation and utility in

conventional pencil-and-paper assessment

∙ When additional description or word information is required

∙ Information search does not refer to open search, but rather

to limited search using closed archives

3Media

Inclusion

Animation, Audio,

Video∙ Presents problem situation using multimedia

4

Tool

Manipulation

/Simulations

Protractor, Ruler,

Calculator, Slider,

Timelapse, Etc.

∙ Using simulated tool and variable manipulation to identify the

relationship between variables (e.g. drawing a diagram using

a protractor and ruler, calculation simulation, virtual experiment)

5 Dialogue Virtual Dialogue

∙ Creates the algorithm of virtual dialogue scenarios that

resemble communications in the real world. (e.g. virtual

messenger)

― [Figure I-3] is an example of an “extended choice” item type using a “hot

spot,” while [Figure I-4] provides an example of “utilizing information,”

item type using a “tab”.

Hot spot – Area selection

*Modified from Item 2 of the 2019 NAEA, middle school Social Studies

[Figure I-3] Example of a “simple presentation” – “extended choice” in Social Studies

Introducing the Computer-Based Test for the National Assessment of Educational Achievement:A Developmental Study for the eNAEA

13

Tab

[Figure I-4] Example of “utilizing information” - “extended writing” in Korean

◈ Development of eNAEA item type application plan and items

○ The study set the test-taker’s response actions and interactional activities as

the two axes and then examined and finalized the item types for eNAEA and

developed an application plan for the test items accounting for the

subject-specific characteristics (refer to <Table I-5>). The basis and the

items of consideration for this plan are outlined below.

<Table I-5> Application plan per eNAEA item type

Interactional

Activities

Test-Taker

Response

1 2 3 4 5

Simple

Presentation

Utilizing

InformationMedia Inclusion

Tool

Manipulation/

Simulations a

Dialogue a

1 Multiple Choice

2 Expanded Choice

3 Connecting Tasks

4 Ordering

5 Short Answer

6 Extended writing

KICE Research Report 2020

14

○ The current assessment framework of each NAEA subject area was reviewed

for their possible inclusion in the computer-based assessment, with a

special focus on the improvement of the competency-based assessment in

the transition to a computer-based assessment.

― “Competencies” set in the test frameworks of Korean, Social Science, and

Sciences subjects were confirmed to be applicable to the computer-based

assessment through an expert committee.

― In Mathematics, “Communication” was newly added to the “Competencies”

set of the test framework, which initially included “calculation and

understanding,” “inference,” “problem-solving,” and “information process.”

― In English, the “Competencies” set of the test framework, which initially

consisted of “English Communication Competency,” was extended by adding

“Community Competency.”

○ The item development process for the eNAEA pilot test (refer to [Figure I-5])

was used to develop the sample items for each category, followed by

detailed discussions on the subject-specific achievement standards, subject

competencies, and implications for item development.

Interactional

Activities

Test-Taker

Response

1 2 3 4 5

Simple

Presentation

Utilizing

InformationMedia Inclusion

Tool

Manipulation/

Simulations a

Dialogue a

7 Correction Tasks

8Completing a

Graph/Figure a

* Short-term (by 2022), Midterm (within 2-3 years after implementation)

** Having two periods in a single cell denotes that there are differences in the introduction and

application within each type.

*** This is based on an average of application periods for the five subjects of eNAEA. There may be

exceptions.

a The item types that can be more affected by the level of technical development.

Introducing the Computer-Based Test for the National Assessment of Educational Achievement:A Developmental Study for the eNAEA

15

3. Analyze the applicability of

technical features to the

2020 NAEA items

2. Develop analytical

framework for eNAEA

items

1. Analyze existing NAEA

items, achievement

standards, and

competency

5. Establish details in the

methods of transition for

existing items and

development process for

new ones

4. Select items from 2020

NAEA to be technically

modified for the pilot

8. Confirm and upload

the item

7. Analyze usability 6. Design methods of

realizing technical

features and assessment

items

Total

Subject-specific / Total Subject-specific / Total Subject-specific / Total

Subject-specific Subject-specific

Total

Subject-specific

[Figure I-5] Item development process for the eNAEA pilot test

◈ eNAEA item development process

○ The study outlines a new test item development process for the eNAEA

based on the analysis of the items of improvement required from the

conventional pencil-and-paper system.

― Compared to the paper-and-pencil version of NAEA, eNAEA requires

additional stages in the test development process: storing items to the item

bank and extracting item from the item bank for test delivery. The study

outlined a new test item development process for the eNAEA, which is

customized to the digital characteristics of the test system, as well as to

the item types utilizing some degree of technology and multi-media

― After analyzing the relative advantages and disadvantages of possible item

development methods, such as “camp-types” and “competition-types,” the

study concluded that both camp-type and competition-type item

development methods should be employed during the initial short-term

stages of eNAEA implementation until the new test becomes stabilized,

after which the ratio of the two methods could be revisited.

KICE Research Report 2020

16

4. Policy Recommendations

○ This study presents its policy recommendations as in [Figure I-6], with the

ultimate goal of utilizing the eNAEA system in a proactive manner to provide

a student- and school-centered student achievement assessment service that

can support innovation in school education and student assessment and

ensure the continuous development of the eNAEA system.

○ Policy suggestions are made from the two points of “supporting the innovation

of school education and student assessment by employing the eNAEA

system” and “supporting the improvement of the eNAEA system.”

― The first point of “supporting the innovation of school education and student

assessment by employing the eNAEA system” includes recommendation to

“provide a school- and student-centered innovative student achievement

assessment service by using the eNAEA system,” “develop and provide a

school-based digital assessment platform (e-assessment platform),” and

“strengthen teachers’ assessment skills (e-assessment expertise) according

to the change of assessment environments.”

― Policy recommendations in the second point of “supporting the improvement

of the eNAEA system” includes the need to “enhance the eNAEA test

development and administration system to ensure test validity and user

experience,” “build and provide a proper infrastructure for the stable

administration of eNAEA,” and “support foundational and in-depth

research to improve the test validity of eNAEA and strengthen the

measurement of student ability and performance.”

Introducing the Computer-Based Test for the National Assessment of Educational Achievement:A Developmental Study for the eNAEA

17

Supporting the innovation of school education and student growth through innovations in assessment

Provide a regular-base diagnostic evaluation service of academic achievement by

integrating the NAEA database and eNAEA system.

Expand the availability of the eNAEA diagnostic evaluation service to strengthen support

toward achievement assessment in the field.

Improve upon the diversity and immediacy of eNAEA assessment results to provide

personalized information to learners.

Carry out a need analysis for the development of the e-assessment platform.

Develop and distribute the e-assessment platform to schools.

Develop teacher training and certification programs to strengthen teacher expertise in

e-assessment.

Utilize teacher learning communities to advertise item types and encourage their use.

Support the enhancement of eNAEA test development tools to ensure test validity.

Support user interface enhancement to ensure user accessibility and experience.

Provide improvement in test administration environment of eNAEA to a school unit.

Provide administrative and financial support to establish tablet-based assessment

systems in line with changes in computer use.

Establish supportive measures toward the needs of special education students in

computer-based assessment.

Conduct research on the effects of the transition to computer-based assessment on

student achievement.

Conduct research on how to examine and use performance data of eNAEA to improve

test validity.

Provide a school- and student-centered innovative student achievement assessment service using the eNAEA system

Develop and provide a school-based digital assessment platform (e-assessment platform)

Strengthen teachers’ assessment skills (e-assessment expertise) according to the change of assessment environments

Enhance the eNAEA test development and administration system to ensure test validity and user experience

Build and provide a proper infrastructure for the stable administration of eNAEA

Support foundational and in-depth research to improve the test validity of eNAEA and strengthen the measurement of student ability and performance

Recommendation

Recommendation

Recommendation

Recommendation

Recommendation

Recommendation

Supporting the innovation of school education and student assessment by employing the eNAEA system

Supporting the improvement of the eNAEA system

Strategy

1

Strategy

2

Strategy

3

Strategy

1

Strategy

2

Strategy

1

Strategy

2

Strategy

1

Strategy

2

Strategy

1

Strategy

2

Strategy

1

Strategy

2

Strategy

3

[Figure I-6] Policy recommendations and strategies

KICE Research Report 2020

18

References

Bridgeman, B. (2009). Experiences from large-scale computer-based testing in

the USA. In F. Scheuermann, & J. Björnsson (Eds.), The Transition to

Computer-Based Assessment (pp. 39-44). Luxembourg: Office for

Official Publications of the European Communities.

Bryant, W. (2017). Developing a strategy for using technology-enhanced items

in large-scale standardized tests. Practical Assessment, Research, and

Evaluation, 22(1), 1-10.

de Klerk, S. (2012). An overview of innovative computer-based testing. In T.

J. H. M. Eggen, & B. P. Veldkamp (Eds.), Psychometrics in Practice at

RCEC (pp. 137-150). Retrieved May 27, 2020, from https://ris.

utwente.nl/ws/portalfiles/portal/5579570/Chapter_12.pdf.

Drasgow, F., & Mattern, K. (2006). New tests and new items: Opportunities and

issues. In D. Bartram, & R. K. Hambleton (Eds.), Computer-Based Testing

and the Internet: Issues and Advances (pp. 59-75). England: John Wiley

& Sons Ltd.

Ministry of Education (March 28, 2019). No child gets left behind in basic

education. Ministry of Education Press Release.

Reich, K., & Petter, C. (2009). eInclusion, eAccessibility and design for all issues

in the context of European computer-based assessment. In F. Scheuermann,

& J. Björnsson (Eds.), The Transition to Computer-Based Assessment (pp.

68-74). Luxembourg: Office for Official Publications of the European

Communities.

Sireci, S. G., & Zenisky, A. L .(2006). Innovative item formats in computer-

based testing: In pursuit of improved construct representation. In S. M.

Downing, & T. M. Haladyna (Eds.), Handbook of Test Development (pp.

329–347). Mahwah, NJ: Lawrence Erlbaum Associates Publishers.

Strain-Seymour, E., Way, W. D., & Dolan, R. P. (2009). Strategies and Processes for

Developing Innovative Items in Large-Scale Assessments. Pearson Education,

Introducing the Computer-Based Test for the National Assessment of Educational Achievement:A Developmental Study for the eNAEA

19

Inc. Retrieved March 29, 2020, from https://images.pearsonassessments.

com/images/tmrs/StrategiesandProcesses forDevelopingInnovativeItems.pdf.

van Lent, G. (2009). Risks and benefits of CBT versus PBT in high-stakes testing:

Introducing key concerns and decision making aspects for educational

authorities. In F. Scheuermann, & J. Björnsson (Eds.), The Transition to

Computer-Based Assessment (pp. 76-84). Luxembourg: Office for Official

Publications of the European Communities.


Recommended