Upload
others
View
0
Download
0
Embed Size (px)
Citation preview
Digitally Signed by: Content manager’s Name DN : CN = Webmaster’s name O = University of Nigeria, Nsukka OU = Innovation Centre
Ugwoke Oluchi C.
Faculty of Education
Department of Science Education,
UNIVERSITY MATRICULATION EXAMINATION SCORES AS PREDICTORS OF FIRST YEAR GRADE POINT AVERAGE
OF STUDENTS IN NIGERIAN UNIVERSITIES IN SOUTH-SOUTH, NIGERIA.
AYAWEI, EREFETEI .K PG//PhD/05/4
2
UNIVERSITY MATRICULATION EXAMINATION SCORES AS PREDICTORS OF FIRST YEAR GRADE POINT AVERAGE
OF STUDENTS IN NIGERIAN UNIVERSITIES IN SOUTH-SOUTH, NIGERIA.
BY
AYAWEI, EREFETEI .K
PG//PhD/05/40059
DEPARTMENT OF SCIENCE EDUCATION, FACULTY OF EDUCATION,
UNIVERSITY OF NIGERIA, NSUKKA
OCTOBER 2013
i
UNIVERSITY MATRICULATION EXAMINATION SCORES AS
PREDICTORS OF FIRST YEAR GRADE POINT AVERAGE OF STUDENTS IN NIGERIAN UNIVERSITIES IN
SOUTH-SOUTH, NIGERIA.
A PhD THESIS SUBMITTED TO THE DEPARTMENT OF SCIENCE EDUCATION,
FACULTY OF EDUCATION, UNIVERSITY OF NIGERIA, NSUKKA
IN PARTIAL FULFILLMENT OF THE REQUIREMENTS FOR THE AWARD OF THE DEGREE OF DOCTOR OF PHILOSOPHY IN
MEASUREMENT & EVALUATION
BY
AYAWEI, EREFETEI .K PG/PhD/05/40059
OCTOBER, 2013
ii
iii
CERTIFICATION
AYAWEI, EREFETEI K, a postgraduate student in the Department of Science
Education with Reg. No. PG/PhD/05/40059 has satisfactorily completed the
requirement for the award of the doctorate Degree in Philosophy of Science Education
(Measurement and Evaluation).
The work embodied in this thesis is original and has not been submitted either in
part or full for any other diploma or degree of this University or any other institution.
_____________________
Erefetei K. Ayawei
Student
iv
DEDICATION
This work is dedicated to the memory of my late parents’ Mr. Joseph Motto & Mrs.
Janet F. J. Ayawei (nee Ekubo) and to the honour of Jehovah God for his mercies as well as all
my HELPERS.
v
ACKNOWLEDGEMENT
With utmost sense of humility, I am most grateful to my supervisor Prof. B.G. Nworgu
who gave me every support I needed. If I had worked at his speed I would have graduated a
long while. He deserves more gratitude than can be expressed here. To Dr. (Mrs.) L.N.
Nworgu, my supervisor’s wife, your words of encouragement literally spurred me to this point
when I was at the verge of abandoning this programme.
To Prof. A. A. Ali, Prof. D.N. Ezeh, Prof. Uchenna Eze, Dr. B.C. Madu, Prof. K.O.
Usman, Dr. A. O. Ovute and Dr. J. J. Agah, Dr. (Mrs.) F.O. Ezeudu, Dr. Christian A. Egbulefu,
Dr. Samuel I. Igbacha, I owe you gratitude. To Prof. F.A. Okwo and Prof. S.A. Ezeudu who
made very useful comments to enrich the content and quality of this work, I am most grateful.
Bro. Obinna Oje, Dr. Davidson Numonde and Chief (Dr.) K. I. Sintei, you were a bundle of
support at every point. To every staff of the Department of Science Education, your labour will
be ever remembered.
I also wish to appreciate Elder (Dr.) Emmanuel O. Denenu & Hon. (Mrs.) Victoria P.
Denenu and every member of the Intercessors Fellowship of Bayelsa (IFB) for their immense
spiritual and moral support. To my father in the Lord, His Lordship, Bishop Ebi S. Belepeigha,
presiding bishop, Global Victory International Ministries, Yenagoa, I am most grateful. I wish
to specially appreciate Bro.Tari MacDonald & Sis. Tari Nyingifa, Dcn. Popo I. & Sis. Ovieni
Popo Jonah, Ch. Freedom & Lady Deborah Etebu, Rev. Segun & Pst. (Mrs.) S. Favour
Balogun all of Global Victory International Ministries, Yenagoa for their support.
To a Mother-in-Israel, Lady (Mrs.) C.N. Nnebedum who made a home for me away from
home, and to the children, particularly Ms. Chinasa F. Ukwueze, thanks for your
encouragement and support.
To Mrs. Yenesom George, former Director of Personnel SUBEB thanks. You were more
than a friend. Mrs. Arangaebi, Director, School Services, Dr. J. Beredugo, Director, Teacher
Training and Mr. Eboh Mazi Amon Chief Admin Officer all of SUBEB, Bayelsa State, I
vi
sincerely appreciate your contribution.
I must pay tribute to my wife, Evang. Mrs. Rhoda E. Ayawei who was more than a pillar of
support all through the programme. To the children: Idimu Emmanuel Idimu, Yeinpatei
Ayawei, Ebiemi Ayawei, Charity Ayawei and Peace Ayawei we can thank God. I also
appreciate the goodwill of Unduemi Yabefa, Angatimi Yabefa, Pretei Yabefa, Ayibamietei
Yabefa and Ebipa Yabefa. Thanks for your support.
To my father, High Chief, Joshua E. Ayawei-Olo, thanks for your encouragement at all
times. You desired to see this day and it has come to pass. To my brothers and sisters: Chief
Godwill & Mrs. Alice G. Ayawei, Bishop Prosper & Rev. (Mrs.) Stella P. Ayawei, Apostle
Afatei Ayawei (Olomiewei I of Koluama) and family, Hon. Nimibofa & Lady (Mrs.) Diepreye
N. Ayawei, Surv. Ebigba & Mrs. Gloria Ebi Ayawei, Mrs. Praise Iniye Ernest (nee Ayawei)
Ebimene Ayawei, Sis. Judith S. Igho (nee Ayawei), Mrs. Suoyo kwomo (nee Ayawei), Ms.
Diepreye Ayawei, Bro. Ebinimi Ayawei, Sis. Emotimi Ayawei, Godday Owika, Mrs. Rose
Tariah (nee Ayawei) and Ms. Florence Ayawei, thanks for placing me on a pinnacle of glory.
God has done it for us.
My gratitude goes to M. Kokobaiye Angaye, former Exams & Records Officer, NDU,
Egberiwarebo Ebilade of Admissions Office, NDU and all other personnel who provided the
relevant data that gave the result to this study.
Finally, I must express my eternal gratitude to Jehovah God for His mercies and the
immeasurable grace to complete this programme. To many more persons’ whose names are not
mentioned here, I appreciate your enormous contribution to my success.
vii
TABLE OF CONTENTS
Title page - - - - - - - - - - - i
Approval page - - - - - - - - - - ii
Certification - - - - - - - - - - iii
Dedication - - - - - - - - - - - - iv
Acknowledgement - -- - - - - - - - - v
Table of content - -- - - - - - - - vii
List of tables - - -- - - - - - - - - xi
List of figures - - - - - - - - - - - - xii
Abstract - - - -- - - - - - - - - xiii
CHAPTER ONE: INTRODUCTION - - - - - - - - 1
Background to the problem - - - - - - - - - 1
Statement of the Problem- - - - - - - - - - 11
Purpose of the Study - - - - - - - - - 12
Significance of the Study- - - - - - - - - - 13
Scope of the Study- - - - - - - - - - - 15
Research Questions- - - - - - - - - - - 16
Hypotheses - - -- -- - - - - - - 16
CHAPTER TWO: LITERATURE REVIEW--- - - - - - 18
A. Conceptual framework - - - - - - - - - - 18
Concept of Unified Tertiary Matriculation Examination (JAMB-UTME) - - - 18
Concept of Post-Universities Matriculation Examinations (PUME) Screening tests - - 19
Concept of test validity - - - - - - - - - 21
Predictive validity - - - - - - - - 25
Procedure for calculating predictive validity - - - - - - 28
Factors affecting predictive validity - - - - - - - - 30
viii
The concept of moderator variables - - - - - - - - 35
Statistical tests for detecting and estimating moderator influence in predictive studies - 39
Statistical implications of moderated.multiple regression analysis- - - - - 42
B. Theoretical Framework - - - - - - - - - - 47
A Point-To-Point Theory - - - - - - - - - 47
Classical Test Theory - - - - - - - - - 48
C. Empirically Related Studies. - - - - - - - - - 49
Studies on predictive validity - - - - - - - - 49
Studies on moderator variables - - - - - - - - - 54
D. Summary of Review of Literature - - - - - - - - - 60
CHAPTER THREE: RESEARCH METHOD - - - - - - 62
Design of the study-- - - - - - - - - - 62
Area of the study-- - - - - - - - - - - 62
Population of the study-- - - - - - - - - - 62
Sample and sampling technique - - - - - - - - 63
Source data collection - - - - - - - - - - 63
Validation of the instrument - - - - - - - - - - 63
Reliability of instrument - - - - - - - - 64
Method of data collection-- - - - - - - - - - 64
Method of data analysis-- - - - - - - - - - - 64
CHAPTER FOUR: RESULTS - - - - - - - - - 70
Research question One - - - - - - - - - 70
Hypothesis One - - - - - - - - - - 71
Research question Two - - - - - - - - - 72
Hypothesis Two - - - - - - - - - - 72
Research question Three - - - - - - - - - 72
Hypothesis Three - - - - - - - - - - 73
ix
Research question Four - - - - - - - - - - 73
Hypothesis Four - - - - - - - - - - 73
Researcher question Five - - - - - - - - - 76
Hypothesis Five - - - - - - - - - - - 80
Research question Six - - - - - - - - - - 80
Hypothesis Six - - - - - - - - - - 87
Research question Seven - - - - - - - - - 88
Hypothesis Seven - - - - - - - - - - 94
Research question eight - - - - - - - - - 95
Hypothesis Eight - - - - - - - - - - - 98
Research question Nine - - - - - - - - 98
Hypothesis Nine - - - - - - - - - - 101
Summary of results - - - - - - - - - 103
CHAPTER FIVE: DISCUSSION OF RESULTS, CONCLUSION, IMPLICATIONS,
RECOMMENDATIONS, AND SUMMARY OF THE STUDY - - - - 104
Discussion of the results - - - - - - - - - - 104
1) The predictive validity of UTME - - - - - - - - 104
2) The predictive validity of PUME - - - - - - - - - 113
3) The predictive validity of the combination of UTME and PUME - - - - 106
4) The influence of gender on the predictive validity of UTME scores - - - - 108
5 The influence of gender on the predictive validity of PUME scores - - - - 109
6) The influence of field of study on the predictive validity of UTME scores - - - 110
7) The influence of field of study on the predictive validity of PUME scores - - - 111
8) The influence of proprietorship on the predictive validity of UTME scores - - 113
9. The influence of proprietorship on the predictive validity of PUME scores - - - 115
x
Conclusion - - - - - - - - - - - 116
Educational implications - - - - - - - - - 119
Limitations of study - - - - - - - - - - 122
Recommendations - - - - - - - - - - 122
Suggestions for further studies - - - - - - - - - 123
Summary of the study - - - - - - - - - 124
References - - - - - - - - - - - 128
APPENDICES - - - - - - - - - - - - 143
Appendix A: List of universities in the south-south geopolitical zone - - - 143
Appendix B: Instrument for data collection- - - - - - - - 144
Appendix C: Letter for access to data - - - - - - - - - 146
Appendix D: Hierarchical Multiple Regression (HMR) output on the combination of
UTME and PUME scores with FYGPA - - - - - 147
Appendix E: HMR output on the influence of gender on UTME scores with FYGPA - - 150
Appendix F: HMR output on the influence of gender on PUME scores with FYGPA - - - 153
Appendix G: HMR output on the influence of field of study on UTME scores with
FYGPA - - - - - - - - - - 156
Appendix H: HMR output on the field of study on PUME scores with regard to
FYGPA - - - - - - - - 162
Appendix I: HMR output on the influence of proprietorship of institution on the
predictive validity of UTME with regard to FYGPA - - - - - 168
Appendix J: HMR output on the influence of proprietorship of institution on
the predictive validity of PUME score with regard to FYGPA - - - 171
Appendix K: Converted UTME and PUME scores - - - - - - - - - 174
Appendix L: C_UTME scores, gender and Interaction term - - - - - - 187
Appendix M: C_PUME scores, gender and Interaction term - - - - - - 199
Appendix N: C_ UTME scores, field of study and Interaction term - - - - 211
Appendix O: C_ PUME scores, field of study and Interaction term - - - - 231
Appendix P: C_ UTME scores, proprietorship and Interaction term - - - - 235
Appendix Q: C_PUME scores, proprietorship and Interaction term - - - - 247
xi
LIST OF TABLES
Table 1: Model summary of hierarchical multiple regression analysis of UTME and PUME
with FYGPA - - - - - - - - 70
Table 2: Summary of coefficients of hierarchical regression analysis for UTME and PUME
with FYGPA - - - - - - - - 71
Table 3: Summary ANOVA of hierarchical regression analysis for UTME and PUME
with FYGPA - - - - - - - - - - - 73
Table 4: Model summary of Hierarchical Regression for X1, G and X1G with FYGPA - 74
Table 5: Summary of coefficients of hierarchical regression analysis for X1, G and X1 * G with FYGPA - - - - - - - - - 74
Table 6: Model summary of Hierarchical Regression for X2, G and X1 G with FYGPA - - 77
Table 7: Summary of coefficients of hierarchical regression analysis for X2, G and X2* G with FYGPA - - - - - - - 78
Table 8: Model summary of Hierarchical Regression for X1, F and X1F with FYGPA - - 80
Table 9: Summary of coefficients of hierarchical regression analysis of
C_X1, F and C_X1F - - - - - - - 81
Table 10: Model summary of Hierarchical Regression for X2, F and X2F with FYGPA - 88
Table 11: Summary of coefficients of Hierarchical regression analysis of X2, F and X2*F - - - - - - - - - - 89
Table 12: Model summary of Hierarchical Regression X1, P and X1P with FYGPA - 95
Table 13: Summary of coefficients of hierarchical regression analysis of X1, P and X1P with FYGPA - - - - - - - - 96
Table 14: Model summary of Hierarchical Regression for X2, P and X2P with FYGPA - 99
Table 15: Hierarchical Regression Analysis coefficients of X2, P and X2P with FYGPA- 99
xii
LIST OF FIGURES
Figure :1 Moderator Model - - - - - - - - 37
Figure 2: Regression lines predicting Ŷ (FYGPA) as a function of X1 (UTME scores) for G (gender) - - - - - - - - - 76
Figure 3: Regression lines predicting Ŷ (FYGPA) as a function of X2 (PUME scores) for gender (G) - - - - - - - - 79
Figure 4: Regression lines predicting Ŷ (FYGPA) as a function of X1 (UTME scores) for field (F0 and F1) - - - - - - - - 84
Figure 5: Regression lines predicting Ŷ (FYGPA) as a function of X1 (UTME scores) for field (F0 and F2) - - - - - - - 85
Figure 6: Regression lines predicting Ŷ (FYGPA) as a function of X1 (UTME scores) for field (F0 and F3) - - - - - - - 86
Figure 7: Regression lines predicting FYGPA as a function of UTME (X1) scores for
field (F0 and F4) - - - - - - - 87
Figure 8: Regression lines predicting Ŷ (FYGPA) as a function of X2 (PUME scores) for
field (F0 and F1) - - - - - - - - - 91
Figure 9: Regression lines predicting Ŷ (FYGPA) as a function of X2 (PUME scores) for field (F0 and F2) - - - - - - - - - 92
Figure 10: Regression lines predicting Ŷ (FYGPA) as a function of X2 (PUME scores) for field (F0 andF3) - - - - - - - - - 93
Figure11: Regression lines predicting Ŷ (FYGPA) as a function of X2 (PUME scores) for field (F0 and F4) - - - - - - - - - 94
Figure 12: Regression lines predicting Ŷ (FYGPA) as a function of X1 (UTME scores) for proprietorship - - - - - - - - - 97
Figure 13: Regression lines predicting Ŷ (FYGPA) as a function of X2 (PUME scores) for proprietorship - - - - - - - - 101
xiii
ABSTRACT
This study investigated the predictive validity of University Matriculation Examination
Scores and the influence of gender, field of study and proprietorship of institution on the
relationship between the matriculation examinations scores and students’ first year grade point
average (FYGPA). Participants in this study consisted of 564 undergraduate (352 male and 212
female) students whose records were obtained from one federal-owned university and one
state-owned university. The study was guided by 9 research questions and 9 hypotheses tested
at 0.05 level of significance. The study adopted a correlational design and collected data from
universities using the Institution and Student Characteristic Pro Forma (ISCP). The data
collected was analyzed using hierarchical multiple linear regression analysis. The major
findings of this study were as follows: (1) UTME scores did not significantly predict FYGPA
(F1, 562 = 1.98, p = .160); UTME scores explained 0.4% of the variance of FYGPA. (2) PUME scores significantly predicted FYGPA (F1, 562 = 194, p = .000); PUME scores also explained 25.8% of the variance of FYGPA. (3) The combination of UTME scores and PUME scores
significantly predicted FYGPA (F2, 562 = 97.34, p = .000); and also explained 25.8% of the variance of FYGPA. (4) The interaction between UTME scores and gender did not
significantly enhance the predictive validity of UTME scores with regard to FYGPA (∆F3, 560 = 3.39, p = .068); and explained 0.6% unique increase in the variance of FYGPA. (5) The
interaction between PUME scores and gender significantly increased the predictive validity of
PUME scores with regard to FYGPA (∆F 3,560 = 8.66, p =.003); and also explained 1.1% unique increase in the variance of FYGPA (6) The interaction between UTME scores and field
of study significantly enhanced the predictive validity of UTME scores with regard to FYGPA
(∆F9,554 = 4.74 p = .001); the interaction explained 3% unique increase in the total variance of FYGPA (7) The interaction between PUME scores and field of study significantly influenced
the predictive validity of PUME scores with regard to FYGPA (∆F9,554 = 10.29, p=.000);the also explained 4.9% unique increase in the variance of FYGPA (8) The interaction between
UTME scores and proprietorship of institution did not significantly enhance the predictive
validity of UTME scores with regard to FYGPA (∆F (3,560) = 1.08, p=.300); the interaction accounted for only 0.2% unique increase in the variance of FYGPA (9) The interaction
between PUME scores and proprietorship of institution significantly enhanced the predictive
validity of PUME scores with regard to FYGPA, (∆F3,560 = 6.17, p=.013); the interaction explained 0.8% unique increase in the variance of FYGPA. One of the implications of this
study is that PUME screening tests serve as a quality control measure in the university system.
Among other things it was recommended that rather than seek for the proscription of the
PUME screening tests, JAMB should redesign the UTME such that it will be have a higher
predictive validity.
.
1
CHAPTER ONE
INTRODUCTION
Background to the study
A test or examination is an assessment intended to measure an individual's level of
knowledge, skill, aptitude, physical fitness, or classification in many other topics (e.g., beliefs)
in a particular area. A test according to Nworgu (2006) is a quantification of the behaviour of a
candidate based on his or her response to the questions set in a particular area. Thus a test score
is a piece of information, usually a number, which conveys the performance of an examinee on
a test. It is a summary of the evidence contained in an examinee's responses to the items of a
test that are related to the construct or constructs being measured. Ogundokun and Adeyemo
(2010) observed that examination marks or scores assigned to a candidate is an indicator of his
or her cognitive achievement in a particular at a given time or over a period of time.
One of the several reasons for developing and administering a test and grading
candidates with test scores is the selection of candidates to fill identified vacant spaces. These
vacant spaces could be for a job or admission of students into institutions of learning.
According to Mehrens and Lehmann (1975) the selection of candidates to fill the vacant places
either in an institution for admission purposes or establishment for the purpose of employment
or the selection of personnel either into the civil service or the military necessitated the conduct
of selection or placement tests. The idea is that, the spaces are fewer than the number of
candidates applying at a particular time and there are specific skills and abilities needed for
success in the expected field. Therefore, selection or placement or entrance examination breeds
stiff competition.
These tests are developed in such a manner that performances in them reflects the
likelihood of success in a similar environment after being exposed to the skills and training in
the envisaged field of engagement. For instance, Grolund (1970) stated that tests used for
placement or selection purposes, performs two basic functions: ascertaining that the candidate
2
possesses the knowledge and skills needed to begin the planned instruction or engagement and
secondly, to ascertain the extent to which the candidate has already mastered the objectives of
the envisaged programme. Thus in order to ascertain that candidates’ possesses the prerequisite
knowledge and skills, universities in Nigeria now uses an applicant's test score on both the
Unified Tertiary matriculation Examination (UTME) and Post Universities Matriculation
Examination (PUME) screening tests as just one of their many admission criteria to determine
if an applicant should be admitted into one of its undergraduate programs. The other criteria in
this case may include the applicant's senior school certificate examination results or its
equivalent. The reason being that those admitted are the most suitably qualified candidates.
The UTME is developed, administered and graded by the Joint Admissions and
Matriculation Board (JAMB). The Board was established, through Decree No.2, 1978 by the
then Federal Military Government of Nigeria, as an agency of the federal-government and an
examining body. Its principal responsibility is to conduct a national matriculation examination
common for candidates seeking admission into all Nigerian universities thereby centralizing
the admission system. This notion is conveyed in the introduction to the Joint Admissions and
Matriculation Board Act 1989:
“… An act to establish the Joint Admissions and Matriculation Board to
administer a centralized admissions system for universities, polytechnics and colleges
of education...” (Joint Admissions and Matriculation Board Act, 1989
Before its establishment and the conduct of UTME, each of the existing universities
conducted her own matriculation examination. The result was that the admission processes
were untidy and uncoordinated. Adamu (1994) and Harbor-Peters (1999) observed that there
were vacant and unfilled spaces in some universities because of multiple offers of admissions,
multiple application and multiple examinations written by students at that time. Nworgu (2006)
observed that the decentralized university admission system created an allowance for
admission irregularities, malpractices and lack of comparability of standards. Afemikhe (2008)
3
also noted that generally there was untidiness and lack of coordination in the admission system
into universities. Similarly, Ukwuije and Orluwene (2010) observed that there were multiple
admissions, poor logistics, and high cost of application as well as travel expenses, tribalism,
regional tendencies, corruption and non-uniformity of standards in the entrance examinations
across different universities. Omodara (2010) stated that there had existed educational
imbalance occasioned by disparity of educational opportunities between the regions in which
the universities were sited and those outside as well as differing admission standards by the
different universities. Uju and Ezeudu (2010) argued that there were several irregularities in the
matriculation examination which were conducted by the respective universities.
While the above where some of the obvious reasons for the establishment of the Board,
the more important reason was the unity of the country and reduction of regional tendencies.
Asein and Lawal (2007) argued that the need to foster national unity among the different ethnic
groups which were recovering from an avoidable 30- month long war necessitated the
establishment of the Board and the centralizing of the admission processes into Nigerian
universities. In summary, the National Policy on Education with regard to universities stated as
follows:
“…admission of students and recruitment of staff into universities and other institutions
of higher learning should be on a broad national basis…” (NPE: 24/26).
The JAMB administered matriculation examinations had several advantages over the
individual universities conducted matriculation examination. First, it provided candidates the
opportunity to obtain a single form and chose an examination centre closest to them without
necessarily travelling to their preferred university for the examination. Secondly, it reduced the
influence of state of origin because candidates had the opportunity of choosing universities far
away from their state of residence. Third, the UTME ensured that all candidates seeking for
admission to tertiary institutions in Nigeria write the same examination thereby ensuring
uniformity and comparability of standards of performance between successive years, zones,
4
states and even candidates (applicants).
However, the validity of the UTME scores which is the correlation coefficient between the
UTME scores and university grades, either the first year Grade point Average (FYGPA) or the
cumulative Grade Point Average (CGPA), has been a concern for users of the UTME and
critics. One of the criticisms of the UTME scores is that they do not adequately predict
individual students’ university grades. For example, Obioma and Salau (2007) observed that
the UTME scores was the least predictor of both the first year grade point average (FYGPA)
and final year cumulative grade point average (CGPA), accounting for 1.5% of the variance in
FYGPA and 0.6% of final CGPA. In a similar study conducted by Ojerinde and Kolo (2009)
the authors observed that UTME scores accounted for a very low amount of variance (2.8%) in
the FGPA.
Others have argued that the scores students obtain in the UTME are not their true scores,
suggesting they may have been influenced by external influences. For instance, Ifedili and
Ifedili (2010) stated that the admission decisions based on UTME scores have created an
access route for malpractice compliant students to be admitted into courses and programmes
they are least qualified for admission. Consequently it has led to a loss of educational equality
and opportunities among applicants and created unfairness in the admission process. Uju and
Ezeudu (2010) also argued that examination malpractice induced by psychological factors such
as fear, stress, insufficient preparation for the examination, low academic ability, laziness
among other things have resulted in seemingly unreliable UME scores. Consequently the
universities have offered admission to unqualified candidates. Omede (2010) stated that
because there seems incongruence between UTME scores and expected course demands. The
author argues that this incongruence exists because the scores candidates obtained in the
examination were induced by external influences that manifested at the time. Consequently
many matriculated student had changed courses which they were originally admitted, to other
courses which seem to be less academically demanding. This is not necessarily due to interest
5
in the new course but necessitated by lack of commensurate ability to match the demands of
the former course or discipline.
These criticisms necessitated the birth of the Post-Universities Matriculation Examination
(PUME) screening tests. Thus the 2005/2006 school year would be long remembered because
it ushered in the PUME screening tests. The PUME otherwise called the Screening Tests are
university-dependent matriculation examination designed and conducted to test the consistency
of performance standards of applicants or candidates who apply for university admission in a
current school year. According to Isine, Deji-Folutile and Amakeodo (2005), the Federal
Government of Nigeria approved the request of the committee of Vice-Chancellors to further
screen candidates to determine the quality of student as reflected in their UTME scores. To
qualify for the PUME, the candidate may have sat for the Unified Tertiary matriculation
Examination (UTME) in the current year and meet a university determined cut-off score/point.
As currently designed the PUME is a screening test, hence different universities adopt different
methods for the screening exercise. In some universities, the screening tests are oral interviews
while in some others it is standardized achievement test. Consequently the criteria for
admission also vary between institutions.
Some universities admit students based on an aggregate of the UTME and PUME
screening test scores. The average performance of the candidates is now assessed in relation to
faculty and departmental cut-off point for possible admission/placement. Proponents see the
PUME screening tests as a great leveller which allows demonstration of consistent individual
ability and a remedial measure that permit universities to make better decisions about those
students for whom UTME scores may not provide wholly consistent information about their
abilities as required for university admission.
However the post-university matriculation examination (PUME) screening tests are not
without its criticisms. First is the concerns of and fear expressed by parents, guardians’ and the
candidates. This group argues that the PUME screening tests will breed favouritism and state
6
of origin bias in university admissions. Some candidates fear that their admission and those of
their loved ones will depend on the strength of the personal relationship between parents,
guardians or relatives with university authorities. The implication is that children of the poor in
the society may not gain university education as they will not have equal access to the
authorities.
Second, the establishment (approval) and subsequent conduct of the PUME screening
tests will lead to lack of uniformity of entrance examination standards which existed before the
commencement of the UTME. Third, some critics believe that apart from breeding issues
relating to state of origin and personal relations in admission decisions, the PUME screening
tests are avenues by university authorities to raise funds both for the institution, cyber café
operators and bank officials. This group contends that there are several in-built charges that
make these forms much more expensive than the UTME. Fourth, there is the opposition from
JAMB official who view the intrusion of the PUME screening tests as unlawful and a subtle
attempt by university authorities to render their employments redundant by performing their
statutory responsibility. For instance, Ifedili and Ifedili (2010) noted that JAMB official
complained bitterly about the PUME screening process as universities usurping their
mandatory roles. For instance JAMB officials argued before the Nigerian Senate that there is
no law mandating the universities to conduct matriculation examination for university
admission.
The dependent or criterion variable was first year grade point average (FYGPA). The
FYGPA of a student is obtained after the first year of study in a particular institution. Most
Nigerian universities uses the 5-point scale ranging from A=5, B=4, C=3, D=2 to E=1 and
F=fail. It is the cumulative grade of the student’s first and second semester examination.
The independent variables (focal predictors) are the candidate’s Unified Tertiary
matriculation examination (UTME) scores and post-universities matriculation examination
(PUME) screening test scores. UTME scores are obtained from candidates who sat for the
7
Unified Tertiary Matriculation Examination conducted by the Joint Admissions and
Matriculation Board. Unlike the Scholastic Aptitude Test (SAT) which is an aptitude test in
which candidates sit for mathematics and verbal aptitude skills, the UTME is a multiple-choice
achievement test with the Use of English compulsory for all candidates and three other subjects
relevant to a candidate’s preferred course of study. The possible total obtainable score of a
candidate is 400 points. On the other hand, PUME scores are obtained by candidates who were
qualified to take the Screening Tests of their preferred universities after meeting a university-
dependent cut-off score. The cut-off score of most federal universities are higher than state
universities. For instance in the 2013/2014 academic year the cut-off score for all federal
universities was pegged at 180 points while state universities cut-off score was 150 points.
These scores qualify the candidate for the Screening Tests and not for university admission.
The structure, form and maximum score of these screening tests are university dependent and
vary from one university to the other. A score in the UTME or PUME screening test is
supposed to be an indicator of the academic ability of an aspiring university undergraduates
and a forecast of what a student can do as well as evidence of the academic background of the
student.
The moderator variables examined in this study are: gender, field of study and
proprietorship of institution because pre-existing subgroups such as these offers easily
identifiable and theoretically meaningful sources of moderator variables. According to Sharma,
Durand and Gur- Arie (1981) a moderator variable is one which systematically modifies either
the form and/or strength of the relationship between a predictor and a criterion variable. It is
believed that properly accounting for these moderator variables in a multiple variables study
will provide for an improved level of prediction and explanation – other than the predictive
validity coefficients that are often reported. Moderator variables are examined because the
classic validation theory used in explaining the predictor – criterion relationship could not
account for complex phenomena especially in cases where the relationship between the
8
criterion variable and predictor variables could be influenced or enhanced by a third variable
called the moderator variable. Specifically, Capoor and Norman (1976) stated that moderator
variables can affect the relationship between the predictor and criterion therefore properly
accounting for their moderating effects can enhance the explanatory power of research analysis
particularly in a multiple regression approach.
Gender in this study is understood to mean the biological and physiological
characteristics that define men and women and not gender in terms of the socially constructed
roles, behaviours, activities, and attributes that a given society considers appropriate for men
and women. With regard to the admission of candidates into Nigerian universities, there are no
different cut-off marks or points for females and males students.
Field of study is here defined as a branch of knowledge that is taught and researched at the
college or university level. The Longman Interactive English Dictionary (2000) defines field as
a department within a university e.g. engineering. Therefore by faculty or field of study means
the branch of study the student chooses to specialize in the university. With regard to
admission into undergraduate programmes in Nigerian universities the cut-off mark is field or
faculty dependent. The fields or faculties of the medical sciences, law and engineering
generally have higher cut-off marks than other faculties or fields of study.
Proprietorship is understood to mean the ownership of the institution. There are two types
of institutions in terms of ownership in Nigeria. These are the public universities and private
universities. Private universities are directly controlled and managed by a non-governmental
organization (e.g. a faith-based organization, trade union, business enterprise or an individual),
or its Governing Board consists mostly of members not selected by a public agency. On the
other hand, public universities are controlled and managed directly by a public education
authority or by a governing body (council, committee etc.), most of whose members are
appointed by a public authority or elected by public franchise. Additionally public universities
in Nigeria are separated into two. These are the federal or state universities. Federal
9
universities are controlled and managed directly by a federal authority or by a governing body
(council, committee etc.), most of whose members are appointed by the Federal Government of
Nigeria. State universities, on the other hand, state universities are controlled and managed
directly by the respective state governments or by a governing body (council, committee etc.),
most of whose members are appointed by the respective state governments.
The federal universities are basically funded by the federal government, while the state
universities are principally funded by the respective state governments with occasional grants
from the Federal Government through her agencies such as the Education Trust Fund (ETF)
and the private universities are owned and funded by private individuals or corporate
organizations. Generally the tuition fees of federal universities are much lower than either the
state or private universities. However, because of the lower tuition fees, they have the largest
student population than either the state or private universities which put undue pressure on the
resources and infrastructure of the respective institutions.
With regard to admission into undergraduate programmes in Nigerian universities the cut-
off mark is university dependent. The federal universities generally have higher cut-off marks
than state universities. For instance in the 2013/2014 school year the cut-off score for all
federal universities was 180 marks whereas the cut-off score for state universities was 150
marks.
Studies exist that provide evidence that moderator variables have influence on the predictive
validity of criterion measures. For instance, Zeidner (1987), JAMB (2007) hypothesized that
the relationship between aptitude test scores and grade point average (GPA) could be
moderated by the age of the candidate taking the test at a particular time. The results of these
studies showed that indeed predictive values differed significantly across different age groups
as hypothesized. Similarly Ojerinde and Kolo (2009) and Ojerinde and Ojo (2009) also tested
the hypothesis that the predictive validity of UTME scores in relation to first year grades could
be moderated by gender. The results showed predictive values higher for male students than
10
female students. Ramist, Lewis and McCamley (1994) showed that women scores were under-
predicted while male SAT scores were over-predicted.
Predictive validity is the extent to which performance on a test is related to later
performance that the test was designed to predict. For example, the UTME and PUME
screening tests are taken by university-bound students to predict their future performance in
college namely, their college GPA. A score in the UTME or PUME screening test is supposed
to be an indicator of the academic ability of an aspiring university undergraduates and a
forecast of what a student can do. If students who scored high on these tests tend to have high
GPAs in colleges, then we can say that the UTME and PUME scores have good predictive
validity. But if there is no significant relation between UTME and PUME scores and college
GPA, then we would say the UTME and PUME scores has low or poor predictive validity,
because it did not predict what it was supposed to. Nworgu (2006) noted that predictive
validity, a form of criterion-related validity provide evidences that make a test instrument
standard. Hence, establishing the predictive validity of the matriculation examination scores is
a means of providing standardization evidences of the examinations. Schmidt and Hunter
(1998) stated that, in terms of practical value, predictive validity is the most important property
of a personnel assessment method because the predictive validity coefficient is directly
proportional to the practical economic value of the assessment method.
A great deal of research has been done on the predictive validity of UTME scores but the
correlations obtained from these predictive validity studies vary greatly and are typically quite
low. However, most of the research on the prediction of first year GPA has not taken into
cognizance the PUME scores as well as the influence of contextual (Moderator) variables.
The fact remains that if the PUME will be used as a selection examination for admission into
Nigerian universities, its predictive power need to be determined. Similarly the influences of
other contextual matters such as gender, field of study and proprietorship of institution on the
11
predictor-criterion relationship need to be examined. This will help the researcher properly
account for the moderating influence of these elements in predictive validity studies.
Since the UTME and PUME are all designed to help in identifying the candidates who
possesses the prerequisite academic ability for success in university programme in Nigerian
universities, the specific measure (dependent variable) used was the first year GPA of the
2006/2007 academic session
Therefore, the goal in this dissertation was to extend the understanding on the predictive
validity of both UTME scores and PUME scores in relation to FYGPA (the criterion variable);
and also to explain the influence of some categorical moderator variables in the relationship
between the criterion and predictors.
Statement of the problem
Candidates seeking admission into Nigerian universities are now faced with the task of sitting
for two compulsory matriculation examinations. These are the Unified Tertiary Matriculation
Examination (UME) and Post-Universities Matriculation Examination (PUME). Candidates are
offered admission based on the average scores of these two examinations according to
departmental/faculty cut-off marks. The persistent poor performance of students who were
admitted into the different programmes in the universities through the nationally conducted
UTME, despite the success, and sometimes, very high scores, necessitated the PUME
screening tests. The observed evidences of low correlation or relationship between the UTME
scores and subsequent outcome of interest – university grades which is represented by the first
year grade point average (FYGPA) became a source of worry and concern to all and sundry. It
becomes vital to examine the predictive strength of the Universities Matriculation Examination
Scores (UMES) that is the UTME scores and PUME scores, with regard to the FYGPA of
students, to ascertain either of the two is a better predictor.
On the other hand, Capour and Norman (1976) observed that moderator variables can
affect the relationship between the predictor and criterion; hence a proper accounting for their
12
moderating effects can enhance the explanatory power of research analysis particularly in a
multiple regression approach. Pre-existing groups such as those based on gender, field of study
and proprietorship of institution offers identifiable and theoretically meaningful sources of
moderator variables.
Some studies have been conducted in the predictive validity of UTME scores and
PUME scores separately. For instance Uju and Ezeudu (2010), JAMB (2007) reported that
there was low correlation between UTME scores and FYGPA, while Owie (2010) observed
that whereas there was a positive correlation between UTME scores and FYGPA and there was
negative correlation between PUME scores and FYGPA. Similarly, Ojerinde and Kolo (2009)
and Ojerinde and Ojo (2009) had observed that contextual variables significantly enhanced the
predictive validity of UTME scores with regard to FYGPA. These contradictory reports and
research results has led to this present study of the predictive validity of UTME scores and
PUME scores with regard to FYGPA as well as the influence of some contextual variables.
Thus framed in the form of a question: are PUME scores better predictors of FYGPA than
UTME scores? What amount of influence do contextual (moderator) variables have on the
relationship between university matriculations examination scores (UMES) and FYGPA?
Purpose of the study
The general purpose of the study was to determine the predictive validity of UTME scores and
PUME test scores on undergraduate students’ performance in Nigerian universities. Specifically,
the study determined:
1. The predictive validity of UTME scores with regard to first year GPA among undergraduates
in Nigerian universities.
2. The predictive validity of PUME scores with regard to first year GPA among undergraduates in
Nigerian universities.
13
3. The predictive validity of the combination of UTME scores and PUME scores with regard to
first year GPA among undergraduates in Nigerian universities.
4. The influence of gender on the predictive validity of UTME scores with regard to first year
GPA among undergraduates in Nigerian universities.
5. The influence of gender on the predictive validity of PUME scores with regard to first year
GPA among undergraduates in Nigerian universities.
6. The influence of field of study on the predictive validity of UTME scores with regard to first
year GPA among undergraduates in Nigerian universities.
7. The influence of field of study on the predictive validity of PUME scores with regard to first
year GPA among undergraduates in Nigerian universities.
8. The influence of proprietorship of institution on the predictive validity of UTME scores with
regard to first year GPA among undergraduates in Nigerian universities.
9. The influence of proprietorship of institution on the predictive validity of PUME scores with
regard to first year GPA among undergraduates in Nigerian universities.
Significance of the Study
This research which examined the predictive validity of University Matriculation
Examination Scores is significant theoretically and practically.
Theoretically, it is significant because theoretical information had revealed that information
with the highest validity seems to have a point-to-point correspondence with the criterion. This
theory explains that the predictive validity of a test is increased to the extent that there is
congruence between measured skills and expected skills. The outcome of the finding is a test of
the validity of the University Matriculation Examinations Scores (UMES) to ascertain if there
is a correspondence between scores obtained at the point of entry and grades at the end of first
14
year of university education.
Furthermore, the addition of theoretical information on moderator variables provides an
increased understanding on the effects or influence of some contextual variables on the
relationship between predictors and criterion.
With respect to the practical significance of this study, the predictive validity of
UTME scores and PUME scores is significant in many respects and would mean different
things to different people. First it is an attempt to increase the knowledge base in the already
existing literature in predictive studies. Predictive studies are significant because they help to
explain the predictive strength of the focal predictor variables (UTME scores and PUME
scores) in relation to the criterion variable (the first year grade point average) which is a
measure of the cognitive abilities of students in Nigerian universities. Furthermore, it has
explained the influence of some categorical variables–namely gender, field or faculty of study
and proprietorship of institution in the relationship between university matriculation
examination scores (UTME scores and PUME scores) and first year grade point average. The
study of contextual variable is important because in some cases the predictive efficiency of an
independent variables and/or form of relationship with a dependent or criterion variable may
vary systematically as a function of another (variable or other variable
Secondly, the findings make contribution to research efforts that are geared or designed
towards improving the UTME and PUME test. Discussion based on the analysis of data
obtained directly from the respective institutions will help the examining authorities (JAMB
and the individual Universities) to reconsider some aspects of either their instrument or the
process of the administration of the test. Additionally, it has provided a comparative analysis of
student performance across institutions in terms of proprietorship, different fields of study and
the gender (sex) of the students. Since this study was a step forward in the traditional predictive
studies in which the relationship between predictor and criterion variable(s) is analyzed, it is
hoped that as the study explores the contextual variables, the interactive effect has provided a
15
greater insight into the complexities that surround the “low predictive phenomenon between
predictor and criterion variables. The analysis and explanation of the outcomes of the data has
also provided an x-ray of the relative strength and weaknesses of the UTME scores and PUME
scores.
In the same vein the result of this study will be useful to the legislators, university
authorities and JAMB. This is so because, the results provided will support the argument for
and against PUME, and for and against the continued relevance of the monopoly of JAMB’s
UME. This is crucial as the continued existence of the PUME has become a debate in the
National Assembly. More so, the universities which have had the misfortune of justifying the
PUME can have empirical data/facts to pursue their case.
Furthermore, future research on the predictive validity of university entrance examination
may also explore several contextual variables such as school climate, teacher quality,
infrastructure in school, number of student in school in relation to study facilities, age of the
institution, proprietorship including private owned universities, etc. in the prediction of college
performances from entry scores.
For parents, their misconceptions about the PUME screening tests as an avenue to enrich
university authorities and a tool for admitting students with close ties to school administrators
will be minimized.
Scope of study
The study covered universities in the South-South geopolitical zone of Nigeria. The study used
students who were admitted into 10 public universities through the selection examinations of
UTME and PUME. They were students who had completed their first year of study
successfully in federal and state universities and those whose records were available for
collection. The South–South has thirteen (13) universities established between 1970 and
2002. Out of these, there are four (4) federal universities, six (6) state universities and three (3)
16
private universities. Furthermore out of the 6 state universities, 3 are specialized universities of
science and technology.
The criterion variable is first year GPA, the focal predictor variables are UTME scores,
PUME scores and a combination UTME and PUME scores, while the moderator variables are
gender, field of study and proprietorship of institution (state university or federal
university).The study examined the predictive validity of UTME and PUME test scores with
regard to first year grade point average of undergraduates in universities in the South-South
zone of Nigeria as well as the influence of categorical variables on the prediction
Research questions.
1. What is the predictive validity of UTME scores with regard to first year GPA of undergraduate
students in Nigerian universities?
2. What is the predictive validity of PUME scores with regard to first year GPA of undergraduate
students in Nigeria universities?
3. What is the predictive validity of the combination of UTME and PUME scores with regard to
first year GPA of undergraduate students in Nigerian universities?
4. What is the influence of gender on the predictive validity of UTME scores with regard to first
year GPA of undergraduate students in Nigerian universities?
5. What is the influence of gender on the predictive validity of PUME scores with regard to first
year GPA of undergraduate students in Nigerian universities?
6. What is the influence of field of study on the predictive validity of UTME scores with regard to
first year GPA of undergraduate students in Nigerian universities?
7. What is the influence of field of study on the predictive validity of PUME scores with regard to
first year GPA of undergraduate students in Nigerian universities?
17
8. What is the influence of proprietorship of institution on the predictive validity of UTME scores
with regard to first year GPA of undergraduate students in Nigerian universities?
9. What is the influence of proprietorship of institution on the predictive validity of PUME scores
with regard to first year GPA of undergraduate students in Nigerian universities?
Hypotheses
This study was guided by the following null hypotheses (Ho) tested at p < 0.05.
1. The predictive validity of UTME scores with regard to first year GPA of undergraduates in
Nigerian universities is not statistically significant.
2. The predictive validity of PUME scores with regard to first year GPA of undergraduates in
Nigerian universities is not statistically significant.
3. The predictive validity of the combination of UTME scores and PUME scores with regard to
first year GPA of undergraduates in Nigerian universities is not statistically significant
4. The predictive validity of UTME scores with regard to first year GPA of undergraduate
students do not differ significantly according to gender.
5. The predictive validity of PUME scores with regard to first year GPA of undergraduate
students do not differ significantly according to gender.
6. The predictive validity of UTME scores with regard to first year GPA of undergraduate
students do not differ significantly according to field of study.
7. The predictive validity of PUME scores with regard to first year GPA of undergraduate
students do not differ significantly according to field of study
8. The predictive validity of UTME scores with regard to first year GPA of undergraduate
students do not differ significantly between federal-owned and state-owned universities.
18
9. The predictive validity of PUME scores with regard to first year GPA of undergraduate
students is not different significantly between federal-owned and state-owned universities
19
CHAPTER TWO
LITERATURE REVIEW
The literature review is arranged under the following sections:
1. Conceptual framework
a. Concept of Unified Tertiary Matriculation Examinations (UTME)
b. Concept of Post-Universities Matriculation Examinations (PUME) Screening tests in
Nigeria.
c. Concept of test validity
d. The concept of predictive validity
e. Procedure for calculating predictive validity
f. Factors affecting predictive validity
g. The concept of moderator variables
h. Statistical tests for detecting and estimating moderator influence in predictive studies
i. Factors affecting the estimate and analysis of the influence of moderator variables.
2. Theoretical Framework
a. Point-To-Point Theory
b. Classical Test Theory
3. Empirically Related Studies.
a. Studies on predictive validity.
b. Studies on moderator variables
c. Studies on UTME scores and PUME scores
Summary of Review of Literature
Conceptual framework
The Concept of Unified Tertiary Matriculation Examination (JAMB-UTME)
The Unified Tertiary Matriculation Examination (UTME), as it is now called, is a
selection achievement test for candidates seeking admission into Nigerian universities. The
20
questions in the different subject areas are developed, administered and graded by the Joint
Admissions and Matriculation Board (JAMB) which was established by Decree No. 2, 1978 by
the then Federal Military Government. Prior to the conduct of the UTME, each of the existing
universities conducted her own entrance or matriculation examination for candidates seeking
admission. Each candidate enters for English Language and three other subjects relevant to
his/her chosen course of study. The total score a candidate is expected to obtain is 400 points
One of the several reasons for the UTME was the need to unify the matriculation
examinations that were being conducted by the different universities at that time. For instance
Adamu (1994) stated that the establishment of the Board was as a result of the initiative of
Nigerian Committee of Vice-Chancellors (CVC) which was worried about multiple
applications for admissions and multiple offers of admissions into Nigerian universities that
was granted to students.
Secondly, and perhaps the most important factor was the need to create national harmony
and cohesion. Asein and Lawal (2007) observed that apart from centralizing the admission
processes, JAMB as one of the post-civil war reconstruction institutions was intended to
provide a platform for students to be admitted into universities in states other than their states
of origin or region thereby fostering the much needed national unity among the tribes in the
country which was recovering from a civil war at the time of its establishment.
The Concept of Post Universities Matriculation Examinations (PUME) screening tests
The Post-Universities Matriculation Examination (PUME) screening tests as the name
suggests are exercises conducted by individual university to further reassess candidates who
chose the particular university during the nationally conducted Unified Tertiary Matriculation
Examination (UTME). Candidates who sit for these screening tests are applicants seeking for
undergraduate admission into the university in the current year of academic activities. The
centre for each PUME examination is centralized within a campus or campuses of the
university. There are certain criteria which a candidate would meet in order to sit for the
21
screening test of the particular university. First, the candidate may have sat for the UTME in
the current application year and obtained a university dependent cut-off score to qualify for
entry into the PUME screening test or exercise (200 for most Federal universities and about
180 for most state universities in the UTME). Secondly, the particular university may have
been chosen as a first or second choice. It is important to note that the PUME screening tests,
for now, are not complementary to the UTME, since candidates not qualified to enter in for the
PUME without the UTME scores. Furthermore candidates are admitted into the preferred
university based on the aggregate performance in the two matriculation examinations.
Obviously, the idea of the PUME is to provide additional information about the commensurate
ability of the student as reflected in the UTME scores to ascertain if the scores the candidate
obtained reflected his/her actual ability. According to Oyedeji (2011) the PUME screening test
is a quality control measure designed by the universities in Nigeria to select the best and most
qualified candidates for admission into Nigerian universities (especially public universities)
and to instill discipline in the university admission process
Similarly, Bassey, Joshua and Joshua (2010) stated that the university administrators seemed
base their argument for introducing the PUME on poor quality of students admitted into the
universities. In other words, the quality of the students produced by the universities is a
function of the quality of the students who were admitted through the matriculation
examination (UME).
According to Ifeanyi (2011) the PUME screening tests provide students the opportunity
to be admitted on merit rather than on special marks obtained from special centre’s and
fraudulent practices. Similarly Lawal, Ossai and Ekundayo (2010) stated that the PUME
screening tests were introduced in the university admission system to raise the quality of
students who seek for admission into the universities because there were discrepancies between
the high UME scores some students obtained and their actual performance upon admission or
enrolment in the university of choice. In the same vein, Owolabi (2009) stated that the PUME
22
was a child of circumstances because of the suspicion that undergraduate students GPA’s of
Nigerian universities did not justify their JAMB/UTME scores. Summing up the arguments,
Obioma and Salau (2007) stated that the justification for the conduct of the Screening exercises
was because there was a consistent low correlation between candidates UTME scores and
subsequent university grades as represented in their CGPA’s.
The concept of test validity
Determining the validity of a test is one of the ways of gathering information from such test to
ensure that the test is accurate. Walk (2005) stated that an understanding of the concept of test
validity is important, because it is the most important criterion, for it determines whether the
test truly measures what it is expected to measure. According to Ebel and Frisbie (1991), the
understanding of the concept of test validity is the foundation for test fairness and the proper
administration and use of tests and other instruments of assessment. Messick (1980) describes
validity as the foundation of testing. That is to say, the quality of a measurement instrument is
determined by its validity. There is a relationship between validity and university entrance
examinations. An entrance examination, which is an instrument for selecting candidates for a
given programme is expected to provide evidence that those successful in the selection
examination or process will do equally well (according to the degree of their success) in the
programme they are admitted. Invariably, the quality of the pre-selection examination is an
evidence of its validity. In other words validity is the evidence of test quality from the
judgment made on the test score obtained at a particular time. For instance, Kane (2008) stated
that the need for the validation of test is derived from legal, scientific and social expectations
that the claims, decisions and judgment to be made based on a (particular) test score justifiable.
Early writers such as Guilford cited in Garson (2008) stated that validity can be equated
with establishing that constructed scales have correlation with a dependent variable in an
intended manner. The author stated that the codification of validity into four types namely:
content validity, construct validity concurrent validity and predictive validity (the latter two
23
being referred to as criterion related validity) came into being in 1954.
However, in recent times there have been strong arguments concerning the unitary theory of
validity as opposed to those who uphold the traditional typology. Within the unitary theory,
there are those who argue that construct validity is the only validity and all other types are
purely dimensions of construct validity in terms of response to the questions raised and
answers given. There are others in the unitary group who believe that content validity is the
only form of validity and all other types are forms of content validity subsumed under it.
Messick (1989) cited in Garson (2008) and Sireci (2007) focused on the interpretation and uses
of test scores and argues that construct validity is the only form of validity with multiple
standards for assessing various components; relevant content, based on sound theory or
rationale, internally consistent body of items, externally correlated with related measures,
generalizable across populations and time and explicit in its social consequences on the test
takers. The authors argued that content validity focuses on labeling constructs; internal validity
focuses on bias on research design and statistical validity focuses on assumptions of meeting
empirical procedures. Furthermore, Ross (2007) also holds the construct-validity unitary
concept. The author argued that historically validity involves three steps which are all
embedded in construct validity, that is, theorizing to develop an hypothesis regarding the
behaviour of a person during a testing situation, gathering data and drawing inference from the
data developed to test the hypothesis.
The other school of thought on the unitary concept of validity focuses on content. For
instance Lissitz and Samuelson in Kane (2008), Ross (2008), Sireci (2008) and Gorin (2008)
argue that content validity is an umbrella term for validity and referred to the other types of
validity (criterion-related validity, theory-based evidence, and evidence on consequence) as
indicators of validity and went further to state that validity is a property of the test independent
of its interpretation or use of the test. Content validity, the authors argues provides operational
definition as indicators of score meaning.
24
However, authors such as Nworgu (2006), Gorin (2008), still hold the traditional
meaning of validity as the extent to which a test measures what it is intended to measure. That
is the test items must measure some intended objectives developed from a hypothesis of the
object measured and the purpose of the test which will provide a basis for judgment about its
validity. For instance, Ferrara in Gorin (2008) placed validity in two terms:
knowledge, skills, processes and strategies (KSPS) about which the test
user would like to make inferences and the observed KSPS which are
applied by the examiners when solving an item (p460).
Some of the traditionalist such as Kane (2008) makes a case for argument based approach to
validation and defined validity as an evaluation of the interpretation and uses of test results.
That in evaluating theory, it is good to operationally define meaning. The author stated that the
argument based approach to validation specifies the inferences and assumptions implicit in an
interpretation or use of test scores and evaluates the possibility of these inferences and
assumptions using appropriate evidence which for or against provides an overall evaluation of
the validity.
In any case, both the content-specific proponent of validity and the construct relevant
authors have all accepted that there are four types of validity, namely: content validity,
construct validity, predictive and concurrent validity (criterion-related validity) though they
emphasized that the criterion-related validity and construct validity are subsumed under
content validity while construct proponents argue that criterion-related validity and content
validity are subsumed under construct validity. For the purpose of this study as Nworgu (2006)
observed there are four types of validity, namely content validity; construct validity
(discriminant validity and convergent validity), predictive validity and concurrent validity
(criterion-related validity).
25
Types of validity
Content validity
According to Cronbach and Meehl (1955) content validity occurs whenever we assess the test
items as representing a sample of universe of items in which the investigator is interested.
Also Nworgu (2006) stated that content validity occurs when the experiment provides adequate
coverage of the subject being studied. This includes measuring the right things as well as
having an adequate sample. Samples should be both large enough and be taken for appropriate
target groups. The perfect question gives a complete measure of all aspects of what is being
investigated. However in practice this is seldom likely, for example a simple addition does not
test the whole of mathematical ability. Content validity is related very closely to good
experimental design. A high content validity question covers more of what is sought..
Construct validity
According to Cronbach (1955), Denga (1987) and Nworgu (2006) construct validity is
involved whenever a test is to be interpreted as a measure of some psychological construct,
attribute or quality that are assumed to be represented in a test performance to explain some
psychological theory. Construct validity occurs when the theoretical constructs of cause and
effect accurately represent the real-world situations they are intended to model. This is related
to how well the experiment is operationalized. A good experiment turns the theory (constructs)
into actual things that can be measured. Sometimes just finding out more about the construct
(which itself must be valid) can be helpful. Construct validity is thus an assessment of the
quality of an instrument or experimental design. It says 'does it measure the construct it is
supposed to measure'. Without construct validity, one is likely to draw incorrect conclusions
from the experiment. Construct validity has two forms: convergent validity and discriminant
validity.
Convergent validity occurs where measures of constructs that are expected to correlate do
so. Convergent validity is the degree to which concepts that should be related theoretically are
26
interrelated in reality.
Discriminant validity occurs where constructs that are expected not to relate do not, such
that it is possible to discriminate between these constructs. Discriminant validity is
the degree to which concepts that should not be related theoretically are not interrelated in
reality. Convergence and discrimination are often demonstrated by correlation of the measures
used within constructs. Convergent validity and discriminant validity together demonstrate
construct validity.
Predictive validity
Cronbach and Meehl (1955), Fitzgerald (1999), and Garson (2008), stated that predictive
validity is intended to predict how a person will perform at a later date on a different of
assessment of his/her abilities using the performance measures of the present. Asher and
Sciarrino (1974) stated that predictive validity is a measure of agreement between results
obtained by an evaluated instrument and results obtained from more direct and objective
measurements, which is often quantified by the correlation coefficient between the two sets of
measurement obtained for the same target population. Therefore, predictive validity is an
attempt to approximate the future performance of an individual using present performance.
Nworgu (2006) argued that predictive validity determines how accurate an instrument of
measurement administered now on a subject or group of subjects can predict a criterion
measurement. Predictive validity is an indicator of the capacity of how a test taken now will
mirror or predict the future performance and behaviour of an examinee or test taker after an
exposure to the future objectives or programme. Some authors have looked at predictive
validity as a determination of statistical properties of a test. For instance, Thorndike and Hagen
(1961) states that predictive validity is primarily an empirical and statistical evaluation of a test
and an indicator of the degree of correlation of the test's score with some chosen criterion
measure of success. Theoretically, the more the correlation, the more effective the particular
test mirrors the criterion that was measured. This idea is also shared by Ahmann and Glock
27
(1958) who stated that an instrument possesses a predictive value to the extent that predictions
of future pupil behaviour made on the basis of the instrument are found to be accurate. This
measures the extent to which a future level of a variable can be predicted from a current
measurement. Also, differential validity and incremental validity have been identified as
two forms of predictive validity.
Differential validity is the outcome of a computed validity coefficient for different
subgroups such as gender, colour, race of examinees in which there are significant variations
among the different subgroups.
Sirenci and Talento-Miller (2006) stated that a test is said to have exhibited differential
validity, if the utility of a set of predictors does not hold up over particular groups of examinees
(e.g., male and female) which ordinarily shouldn't be so. Linn (1978, 1982, and 1984) defined
differential validity as the differences in the size of the correlation or validity coefficients for
different groups of examinees. This is measured by fitting separate regression lines for each
group and then testing for statistically significant difference between the slopes of intercept.
Also Young (2001) explained that differential validity refers to the situation where a test
is predictive for all groups but to different levels of validities. In other words, the test lacks
equivalent validities for the different groups of examinees. The question about differential
validity is, therefore, a question about whether the correlation between the predictors and the
criterion are different for the different identifiable groups of examinees. The groups in this
respect could be gender, faculty of study, type of institution, age of the establishment of the
institution, etc. On the other hand, differential prediction refers to the situation where the
derived prediction equations for the different subgroups of examinees are significantly
different from one group to another group. The question about differential prediction,
therefore, is a question that relates to whether the predictions models obtained for the different
subgroups of examinees is different. Linn (1978, 1982), Shepard (1987) stated that differential
prediction is determined by comparing regression systems for equality of regression slopes and
28
equality of regression intercepts in the respective prediction equations across the groups of
interest.
Linn (1982) stated that both differential validity and differential prediction are important in
the validation of the scores on a test. However, differential prediction is the more crucial
because differences in prediction have more relevance when considering fairness in selection
than do differential validities. In any case, research on both differential validity and differential
prediction are important because they consistently show evidence of differences in the
prediction equations across subgroups of examinees On the other hand, Kyei-Blankson (2005)
defined differential prediction as a statistical research outcome where the prediction equation
and standard of errors of estimate obtained from analyses have significant variation among
different subgroup of examinees. That is to say, a single prediction equation will not be useful
for the different subgroup of examinees.
Incremental validity on the other hand refers to the degree in which a test improves on
decisions that can be made from existing information, such as the base rate of the attribute
being measured and other measures that are available. The term is also used to denote the
validity of a test arising from successive refinements or improvements of it in the light of
experience or evidence, usually including the removal or replacement of test items that turn out
to be problematic or that yield different results from the others.
For example, Bridgeman (2002) while undertaking a meta-analysis of the writing tests of
the SAT I found out that in four colleges in New Jersey, the incremental validity of writing
tests to the prediction of FYGPA or English course grades ranged from 0.004 to 0.016 for the
four colleges. Furthermore, Bridgeman and Lewis (1991) established the fact that writing tests
had greater predictive value in courses that were more likely to use essays in the determination
of course grades, than others, which did not necessarily use essays in the determination of
course, grades such as in the engineering.
29
Procedure for calculating predictive validity
Over the years, res