30
Using Quick Assessment Methods to Generate Meaningful Data from One-Shot Instruction Sessions Laura W. Gariepy Virginia Commonwealth University Photo courtesy of Microsoft Office Clip Art

2011 Virginia Library Association Conference

Embed Size (px)

DESCRIPTION

"Using Quick Classroom Assessment Methods to Generate Meaningful Data from One-Shot Instruction Sessions." In retrospect, probably should have titled this something along the lines of "Using Rubric-based Assessment in One-Shot Instructions Sessions."

Citation preview

Page 1: 2011 Virginia Library Association Conference

Laura W. Gariepy Virginia Commonwealth University

Using Quick Assessment Methods to Generate Meaningful Data from One-Shot Instruction Sessions

Laura W. Gariepy Virginia Commonwealth University

Photo courtesy of Microsoft Office Clip Art

Page 2: 2011 Virginia Library Association Conference

A little about the classes I teach…

Photo courtesy of flickr user taberandrew via a Creative Commons license

Page 3: 2011 Virginia Library Association Conference

Overview • Developed exercise(s) to assess student

learning based on specific learning outcomes

• Developed analytic rubrics to assess student responses (Wiggins, 1998)

• Administered exercises to students before, during, or after classes

• Assessed student responses (ongoing)

Photo courtesy of flickr user pinksherbet via a Creative Commons license

Page 4: 2011 Virginia Library Association Conference

Why this approach? • Authentic assessment: assesses

artifacts of learning • Rubrics utilize descriptive criteria to

assess complex student answers in manageable chunks (Oakleaf, 2009)

• Mutually beneficial for students and for me

Photo courtesy of flickr user pinksherbet via a Creative Commons license

Page 5: 2011 Virginia Library Association Conference

Students experience

active learning

opportunity; we get the

information we’re after

Photo courtesy of flickr user tamar weinberg via a Creative Commons license

Page 6: 2011 Virginia Library Association Conference

Plug: Megan Oakleaf and RAILS

• Assistant Professor at iSchool at Syracuse University

• Information Literacy Instruction and Assessment

• Founded Rubric Assessment of Information Literacy Skills (RAILS): www.railsontrack.info

• Suggested reading included in packet

Page 7: 2011 Virginia Library Association Conference

Exercise One: Turning your Research Question into a Search

Strategy

Photo courtesy of flickr user pinksherbet via a Creative Commons license

Page 8: 2011 Virginia Library Association Conference

Students will be able to develop a topic-relevant search strategy related to a research

question in order to search library resources effectively.

Learning Outcome:

Photo courtesy of flickr user pinksherbet via a Creative Commons license

Page 9: 2011 Virginia Library Association Conference
Page 10: 2011 Virginia Library Association Conference
Page 11: 2011 Virginia Library Association Conference

Don’t strain your eyes! Refer to packet to read full rubric.

Page 12: 2011 Virginia Library Association Conference

Photo courtesy of flickr user the.sprouts via a Creative Commons license

Page 13: 2011 Virginia Library Association Conference

Exercise Two: Topic Discipline Database Article

(TDDA)

Photo courtesy of flickr user pinksherbet via a Creative Commons license

Page 14: 2011 Virginia Library Association Conference

Students will be able to distinguish between multidisciplinary and discipline specific databases in

order to select most appropriate resources and maximize relevant

results.

Learning Outcome One:

Photo courtesy of flickr user pinksherbet via a Creative Commons license

Page 15: 2011 Virginia Library Association Conference

Students will be able to locate articles in library databases in order to

investigate and refine a research question.

Learning Outcome Two:

Photo courtesy of flickr user pinksherbet via a Creative Commons license

Page 16: 2011 Virginia Library Association Conference
Page 17: 2011 Virginia Library Association Conference
Page 18: 2011 Virginia Library Association Conference

Don’t strain your eyes! Refer to packet to read full rubric.

Page 19: 2011 Virginia Library Association Conference

Photo courtesy of flickr user the.sprouts via a Creative Commons license

Page 20: 2011 Virginia Library Association Conference

The results are in! (Well, some of them).

Image courtesy of flickr user blprnt_van via a Creative Commons license

Page 21: 2011 Virginia Library Association Conference

- Assessment complete for 10 sections of UNIV 200 (about half)

- 227 students had the opportunity to complete and turn in both exercises

- 176 (78%) turned in Search Strategy Exercise - 177 (78%) turned in TDDA Exercise

Image courtesy of flickr user blprnt_van via a Creative Commons license; modified by lwgariepy

Page 22: 2011 Virginia Library Association Conference

- Average scores were significantly higher for TDDA than Search Strategy: - 60% avg score for Search Strategy

- 78% avg score for TDDA

Image courtesy of flickr user blprnt_van via a Creative Commons license; modified by lwgariepy

Page 23: 2011 Virginia Library Association Conference

- Differences in exercises completed in class versus either before or after class:

- Big difference in percent of exercises completed

- Almost no difference in average scores

Image courtesy of flickr user blprnt_van via a Creative Commons license; modified by lwgariepy

Page 24: 2011 Virginia Library Association Conference

- Search Strategy Exercise: In class vs. Outside of class - Percentage turned in:

- In class: 95% - Outside (before) class: 65%

- Differences in average scores: - In class: 61% - Outside (before) class: 60%

Image courtesy of flickr user blprnt_van via a Creative Commons license; modified by lwgariepy

Page 25: 2011 Virginia Library Association Conference

- TDDA Exercise: In class vs. Outside of class - Percentage turned in:

- In class: 91% - Outside class (homework): 59%

- Differences in average scores: - In class: 78% - Outside class (homework): 76%

Image courtesy of flickr user blprnt_van via a Creative Commons license; modified by lwgariepy

Page 26: 2011 Virginia Library Association Conference

- Problem Spots - Search Strategy Exercise:

- Criteria D: Truncation - Average score of only 21%

- TDDA Exercise: - Criteria C: Discipline-specific Databases - Average score of 61%

Image courtesy of flickr user blprnt_van via a Creative Commons license; modified by lwgariepy

Page 27: 2011 Virginia Library Association Conference

Benefits

• Exercises encouraged student learning

• Rubrics are good for assessing complex answers as objectively as possible

• Flexible • Rich data

Photo courtesy of Microsoft Office Clip Art

Page 28: 2011 Virginia Library Association Conference

Challenges:

• Time • Articulating

everything • May or may not

focus on retention

• In one-shots, you lose some of the advantages of rubrics

Image courtesy of flickr user seandalai via a Creative Commons license

Page 29: 2011 Virginia Library Association Conference

Down the Road: Future Uses and Applications (for me and for you!)

Photo courtesy of Microsoft Office Clip Art

Page 30: 2011 Virginia Library Association Conference

Classroom Assessment

+ Teaching / Learning

Questions? Questions?

Laura W. Gariepy [email protected]

(For references, visit

http://bit.ly/VLA2011gariepy)

Image courtesy of flickr user woodleywonderworks via a Creative Commons license