Upload
fabio-r-arico
View
120
Download
1
Embed Size (px)
DESCRIPTION
A description of the teaching methodology used to raise student confidence and assess its relation to student engagement and student attainment.
Citation preview
1
When Student Confidence Clicks
Academic Self-Efficacy
and Learning in HE
Fabio R. Aricò [email protected]
Dec 2013
ACKNOWLEDGEMENTS
UEA-HEFCE Widening Participation Teaching Fellowship
HEA – Teaching Development Grant Scheme
2
OUTLINE
1. Motivation student experience vs. self-efficacy
2. Teaching protocol embedding self-efficacy via SRS
3. Preliminary results learning, self-efficacy, satisfaction
4. Summary feedback welcome!
3
MOTIVATION
Are my answers correct? I’m so confused…
Is this going to be in the exam? Are you sure?
But what if money supply contracts rather than increasing?
Yes, we checked them together already.
Yes, we spoke about it in class and practiced.
You know how to do the reverse, you showed me. Relax.
4
MOTIVATION
Typical problems analysed in recent pedagogic literature:
• Students may encounter difficulties with the course materialsupport sessions, office hours, targeted support interventions.
• Students may display low levels of engagementrevision of the curriculum, innovations in teaching, teaching technologies, partnership lecturer-students.
5
MOTIVATION
Additional problem:
• Students may experience low confidence levelsanxiety over preparation;peer-pressure and competition;inability to self-assess and detect problems.
• The recent changes in HE practice exacerbate this problemthe ‘student experience’ model targets support and satisfaction;students run the risk of being put ‘at the heart of the system’ aspassive receivers, rather than confident owners, of their learning.
6
REACTION
Re-visit the concept of Academic Self-Efficacy:
students’ confidence in their ability to accomplish specific academic tasks or attain specific academic goals (Bandura, 1997).
Teach students how to become confident and independent learnershelp them to self-assess and diagnose problems;enable them to seek appropriate forms of support;increase the rate of retention of widening access students;enhance employability skills all along the academic journey.
7
REACTION in practice
Develop a teaching protocol embedding Academic Self-Efficacy as an independent learning outcome, parallel to the curriculum.
Stage 1: Investigation and assessment of student Self-Efficacy- experiment with Student Response Systems (clickers);- explore correlation between attainment and confidence.
Stage 2: Extension of dataset (add student record data)Extension to qualitative analysis (e.g. focus groups and interview)Targeted intervention to increase Self-Efficacy levels.
8
TEACHING PROTOCOL – the module
Introductory Macroeconomics Level 1 – compulsory year-long module - 170 students
Lectures traditional frontal-teaching (10 per sem.)
Seminars small group, pre-assigned problem sets (4 per sem.)
Workshops large group, problem-solving sessions (4 per sem.)
Support Sessions non-compulsory drop-in sessions (4 per sem.)
9
TEACHING PROTOCOL – clicker technology
10
TEACHING PROTOCOL – the innovations
Lectures interaction via clicker technology
Seminars revision questions + understanding questions
Workshops closing questions:
was the lecture enjoyable?
was the material difficult?
Support Sessions online report of clicking session + feedback
11
TEACHING PROTOCOL – the innovations
Seminars preliminary Seminar Quizzes (paper-based)
Seminars 3 revision/understanding questions
Workshops 2 confidence/self-assessment questions
Sessions open-answer comments
Support Sessions online report of Seminar Quiz
- solutions and overall performance- individual performance available- response to open-answer comments
12
TEACHING PROTOCOL – the innovations
Extra-Curricular Activities to promote engagement and Self-Efficacy
Seminars Module Facebook Page + Blackboard pages
- ‘challenges’ to encourage further study- interaction and participation
Seminars Voluntary in-lecture presentations (5 minutes)
- to exploit demonstration effects
Support Sessions Campus Vouchers (for engagement, not attainment)
13
TEACHING PROTOCOL – the innovations
Workshops peer-instructed flipped classroom approach
Seminars standard algorithm:
1. Quiz questions + Confidence questions (no solution)2. Peer-instruction learning3. Quiz questions + solutions4. Problem-set questions4. Feedback questions:
- what was the cause of mistakes/problems?- did you enjoy using clickers?- were clickers useful to your learning?
Support Sessions online report of clicking session + feedback
14
TEACHING PROTOCOL – the methodology
Focus attainment, engagement, academic self-efficacy role of the SRS (clicker) technology
Learning analytics rich dataset = clicker and paper-based responsesSeminars matched demographics from student records
uncover correlation patterns
Qualitative data focus group and individual interviewsSessions feedback from students
Support Sessions provide the narrative to interpret the analytics
15
An example using data from the 2012-13cohort
16
EXAMPLE – description
• In the following example I present a comparison of two workshop sessions that have taken please over the academic year 2012-13 over Week 5 and Week 9 of the Autumn Semester.
• The data for the academic year 2013-14 is not yet available for presentation purposes, but displays quite different patterns.
• Please regard this example as a description of the methodology more than a discussion of findings.
17
EXAMPLE – workshop structure
1. Set of quiz questions – collect responses via clickers.Students can see the distribution of answers, but no solution given.
2. Ask students to rate their confidence in giving a correct answer (question by question) – collect responses via clickers.
3. Allow for 20mins discussion on quiz questions.
4. Ask the same questions – collect responses via clickersProvide a solution and a discussion to each question.
18
EXAMPLE – dataset
2 workshop sessions: Week 5 and Week 9 – Autumn Semester
attainment number of correct responses per question/per student
confidence “How confident do you feel about the answer
given to Question X?” [4 levels]
learning “I feel that the clickers technology has contributed to my learning experience in today’s workshop” [4 levels]
satisfaction “I enjoyed using clickers in today’s workshop” [4 levels]
Collapsed 4-level responses into dummy response [0,1].
19
EXAMPLE – dataset
Student Q1 Q2 Q3 …
1 0 1 1
2 1 0 0
3 1 1 …
…
performance per question
pe
rform
ance
pe
r stud
en
t
20
EXAMPLE – attainment by question
Week 5
% correct responses■ 1st round ■ 2nd round
21
EXAMPLE – attainment by student
Week 5
% correct responses 1st round
% co
rrect respo
nses 2
nd
rou
nd
22
EXAMPLE – student confidence by question
Week 5
■ % 1st correct responses ■ % confident responses
23
EXAMPLE – student confidence by student
Week 5
% co
nfid
ent resp
on
ses
% correct responses 1st round
24
EXAMPLE – Comparing Week 5 to Week 9
initial preparation
final learning outcome
peer-instruction effect
self-efficacy indicator
problem set difficulty indicator
student self-assessment skills
25
EXAMPLE – Comparing Week 5 to Week 9
Week 9Week 5
more of a learning/preparation problem than a confidence problem!
26
EXAMPLE – Student perception of SRS
Student satisfaction
91% 56%
Student learning perception
88% 73%
27
SUMMARY
• Teaching protocol with interventions to assess/enhance Academic Self-Efficacy and self-assessment skills.
• Mixed-methods approach to disentangle the relationship between engagement, attainment, and academic self-efficacy using student demographics.
• Assessment on the role of SRS technology (clickers) in promoting ASE.(A research question not yet covered in related literature).
28
FEEDBACK NEEDED
• More effort needed to map this project into the literature on ASE.
Academic references? Resources? Interesting readings?
• Explore the role of ASE questionnaires to facilitate comparisons with related studies and validate results.
Which additional instruments could be deemed as useful?
• Further opportunities for dissemination and discussion.
I would be glad to engage with these!
29