4
this issue Open Day P.1 Psychometrics 101 P.2 Psychometrics 101 P.3 New paper P.4 Goodbye P.4 Staff News P.4 Just for Fun P.4 Facebook followers Twier followers CETAP News Internal staff newsleer for CETAP April 2017 UCT Open Day CETAP staff were present at UCT’s Open Day on Saturday, 22nd April. Visitors had their NBT quesons answered at our stall on University Avenue, and Emlyn Balarin and Ja- nine Dunlop gave three presenta- ons about the NBTs over the course of the day. NBT-by-numbers 1733 273

ETAP News - University of Cape Town · ETAP News Internal staff newsletter for ETAP April 2017 UT Open Day ETAP staff were present at UT’s Open Day on Saturday, 22nd April. Visitors

  • Upload
    others

  • View
    9

  • Download
    0

Embed Size (px)

Citation preview

Page 1: ETAP News - University of Cape Town · ETAP News Internal staff newsletter for ETAP April 2017 UT Open Day ETAP staff were present at UT’s Open Day on Saturday, 22nd April. Visitors

this issue

Open Day P.1 Psychometrics 101 P.2 Psychometrics 101 P.3

New paper P.4 Goodbye P.4

Staff News P.4 Just for Fun P.4

Facebook followers

Twitter followers

CETAP News Internal staff newsletter for CETAP

April 2017

UCT Open Day

CETAP staff were present at UCT’s

Open Day on Saturday, 22nd April.

Visitors had their NBT questions

answered at our stall on University

Avenue, and Emlyn Balarin and Ja-

nine Dunlop gave three presenta-

tions about the NBTs over the

course of the day.

NBT-by-numbers

1733

273

Page 2: ETAP News - University of Cape Town · ETAP News Internal staff newsletter for ETAP April 2017 UT Open Day ETAP staff were present at UT’s Open Day on Saturday, 22nd April. Visitors

Psychometrics 101 Natalie le Roux

Part 5b: Achievement tests In the final session of this series, the focus is on how test developers apply the principles of psychometrics in achievement tests. This will be a brief introduction to the scoring, analysis and score interpretation of large-scale achievement tests. Unlike in psychological tests, in achievement tests there is always a correct or partially correct answer to each item. An example of an item with only one cor-rect answer would be: “What is the legislative capital city of South Africa?”. An example of an item for which partial credit could be awarded based on the number of correct answers given would be: “Which country has three capital cities? Name the country and the cities.” In South Africa, there are two national tests that are currently used for first time entry to higher education, namely the National Benchmark Tests (NBTs) and the National Senior Certificate Examination (matric). Once they have been developed (undergone the rigorous process of test development) and administered to the test writers, the completed tests or examination papers are ready to be scored. In the case of the matric examination papers, these are marked by experienced grade 12 subject teachers. For the NBTs, the answer sheets are read by electronic scanners. Once the results (raw data) of the tests are verified and checked and sometimes rechecked, the data team will run item analysis and test analysis on the raw data. The above is a general guide on how the process is undertaken and each examination board or test development agency have clear and specific guidelines on how they score and analyse their data. Analysis Classical Test Theory (CTT) and Item Response Theory (IRT) models are used to perform item analysis. Cast your mind back to session 4 of this series, where the concept of CTT and IRT was introduced. The raw data is analysed for the whole test and for each individual item. Item analysis is a process that assesses the quality of the individual items and the test as a whole (DeMars, 2010; Polit, & Beck, 2012). Summary statistics as seen in table 1 are generated. Table 1: Item analysis information

IRT models generate similar information and more information depending on the IRT that is used. There are sophisticated computer software packages that are currently in use to analyse the raw data, such as Parscale, Multilog, Bilog, RUMM and many more. See box 1 for an example of typical reliability output generated in the Statistical Package for the Social Sciences (SPSS).

Summary statistics - such as mean, highest and

lowest score and more.

Reliability of the test as a whole

(internal consistency of the test

Item difficulty -proportion of writers who an-

swered item correct

Point biserial correlation - the correla-

tion between item and the total score on

the test of a writer

Discrimination index – the difference between

low performing and high performing test writers

who correctly answered the item

Box 1: The reliability index for a 11-item test. The Cronbach’s alpha (α) of .741

indicates the internal consistency or the 11-item test. If item bss01 is removed

from the test the reliability decreases from .741 to .690.

Box 2 is an example of an item response function for a polytomous (partial credit). On the x-axis

is the theta (ability of the writer) and on the y-axis, is the probability of a correct response. This

item has 3 response categories (incorrect, partially correct and completely correct).

Page 3: ETAP News - University of Cape Town · ETAP News Internal staff newsletter for ETAP April 2017 UT Open Day ETAP staff were present at UT’s Open Day on Saturday, 22nd April. Visitors

Score interpretation

It is important to mention the difference between criterion- and norm-referenced

tests. The NBTs are an example of criterion-referenced tests since they measure the

test writer’s performance against a set of criteria (e.g. academic literacy). The ma-

tric examination, on the other hand, is an example of a norm-referenced test since

it ranks a test writer in relation to the other test writers in relation to what they are

being assessed on (e.g. life sciences). This score also gives an indication of how well

the test writer has mastered the subject matter. The NBTs assess the readiness of a

test writer to undertake higher education and therefore the test writer’s score gives

an indication of the readiness of the test writer. It also highlights what the test writ-

er will be able to do and what challenges he/she might face with the academic de-

mands of higher education.

In the next months, you can look forward to reading more about testing and perti-

nent issues around testing: What is the value of online testing, what is computer-

ised adaptive testing, international comparative assessments, high stakes testing

and how to combat cheating just to name a few.

References and sources

DeMars, C. (2010). Item response theory: Understanding statistics measurement. Oxford: Oxford University Press.

Source for Graph 1: http://psych.hanover.edu/classes/ResearchMethods/Assignments/reliability-1.html

Source for Graph 2: U.S. Department of Education, Institute of Education Sciences, National Center for Education Statistics, National

Assessment of Educational Progress (NAEP), 2000 and 2001 Assessment.

Source for Graph 3: https://www.princetonreview.com/college-advice/good-psat-score

Polit, D. F., & Beck, C. T. (2012). Nursing research: Generating and assessing evidence for nursing practice (9th ed.). Philadelphia:

Wolters Kluwer Health/Lippincott Williams & Wilkins.

Carol Bohlmann, Robert Prince, and Andrew Deacon have had a paper published in the journal, Pythagoras:

BOHLMANN, C., PRINCE, R., DEACON, A.. Mathematical errors made by high performing candidates writing the National

Benchmark Tests. Pythagoras, 38, Apr. 2017. Available at: <http://www.pythagoras.org.za/index.php/pythagoras/

article/view/292>.

Psychometrics Part 5b: continued

Box 3: An example of a score report on the Preliminary Standardised

Achievement test (PSAT). This score report indicates the test writer

need to strengthen their reading and writing skills.

New paper published

Page 4: ETAP News - University of Cape Town · ETAP News Internal staff newsletter for ETAP April 2017 UT Open Day ETAP staff were present at UT’s Open Day on Saturday, 22nd April. Visitors

Staff News

Centre for Educational Testing for Ac-

cess and Placement

Level 4, Hoerikwaggo Building

University of Cape Town

www.cetap.uct.ac.za

Telephone: +27 (0)21 650 5462

Fax: +27 (0)21 650 5466

Goodbye

CETAP said a sad “goodbye” to project assistant, Yusrie

Ismail, whose contract ended at the end of April. His

send-off included some warm words from Estelle Mur-

ray, a goodbye gift from the staff, and a companiona-

ble lunch.

Send submissions to:

[email protected]

by the 20th of each month.

Just for Fun

Naomi Msusa gave birth

to a healthy baby boy,

Jaydon Msusa, on the

11th of April.