27
JUSP: building on success Angela Conyers and Jo Lambert 22 nd July 2013, Northumbria International Conference

Summary

  • Upload
    naida

  • View
    38

  • Download
    0

Embed Size (px)

DESCRIPTION

JUSP: building on success Angela Conyers and Jo Lambert 22 nd July 2013, Northumbria International Conference. Summary. JUSP introduction Interoperability with other services Increased functionality Usage profiling Extending beyond e-journals E-book trial JUSPConsult Next steps. - PowerPoint PPT Presentation

Citation preview

Page 1: Summary

JUSP: building on successAngela Conyers and Jo Lambert22nd July 2013, Northumbria International Conference

Page 2: Summary

Summary

• JUSP introduction• Interoperability with

other services• Increased functionality

• Usage profiling

• Extending beyond e-journals

• E-book trial

• JUSPConsult• Next steps

Page 3: Summary

Introduction and aims

• Supports libraries by providing a single point of access to e-journal usage data

• Assists management of e-journals collections to inform evaluation and decision-making processes

• Enables usage comparisons and trend analysis

Page 4: Summary

Who?

Page 5: Summary
Page 6: Summary

Partnership – libraries and publishers

• Delivering a service to the community

• Collaboration and co-development in conjunction with libraries

• Working with publishers to provide enhanced customer service

Page 7: Summary

Standards and interoperability

• Standards– COUNTER– SUSHI

• Interoperability– KB+– ERM systems– RAPTOR-JUse

Page 8: Summary

KB+

Page 9: Summary

RAPTOR-Juse: a Jisc funded project

Open source software suite for accounting of authentication information, primarily designed to assist organisations account for e-resource usage. Who is using the resources?

The Journal Usage Statistics Portal (JUSP) provides a "one-stop shop" for libraries to view, download and analyse their e-journal usage reports from participating publishers. What are they using?

Page 10: Summary

Technical challenges

No common identifier for the ‘service’ that provides access to the resource. • RAPTOR records this as ‘entityID’• COUNTER reports harvested by

JUSP are at platform or publisher level.

Combining daily and monthly figures• RAPTOR is event-based, reporting

on authentication events • JUSP reports on aggregated

information relating to monthly article requests for journals.

Page 11: Summary

Data from RAPTOR and JUSP displayed together in RAPTOR

• JUSP JR1/JR1a reports by journal platform

• RAPTOR report showing use of one resource by School

Page 12: Summary

True, false or maybe?

School A has 65% of authentication events for an ‘entityID’ in RAPTORTherefore School A has 65% of the usage of the platform or publisher in JUSPIs this a safe way of allocating budgets?Are we comparing apples and pears? Some caveats•RAPTOR services contain more content than JUSP platforms or publishers – databases, ebooks etc. •JUSP data are presented at platform/publisher level – need to go to JUSP to get to title level

Page 13: Summary

What has been learned?

• Great interest from the community in the idea behind the RAPTOR-JUse project of combining authentication and usage data

• Great deal has been learned about the technical processes involved in combining these two sets of data

• Is RAPTOR-JUse the way forward, or is the jury still out on how best to discover who are the actual users of our e-resources?

Page 14: Summary

Usage profiling

• How well are we doing?

• How does our usage compare with others?

• What constitutes good usage?

Page 15: Summary

How does it work?

Individual libraries can view their own usage against averages for those in the same:• Jisc band• Regional group (as defined by HESA)• University or library group (Russell group, 1994, post

-92 universities etc) • Average number of FTE users in each of the

appropriate groups

Page 16: Summary

Usage profiling report

Page 17: Summary

What is not included

• Means of identifying individual librariesLibraries only have access to their own data to compare with a set of averages

• CostsJUSP does not contain any cost data for individual libraries

• Deal informationAverages are based on total JR1 and JR1a requests for each publisher, irrespective of deal or collection taken.

Page 18: Summary

How will it be used?

• To use when comparing budgets in other institutions with actual usage

• To monitor effect of promotional campaigns

• To view trends across years

• To help provide context for what constitutes ‘good usage’

• To analyse reasons for usage that was higher or lower than average

Page 19: Summary

E-book trial

Small pilot trial with 2 publishers and members of the JUSP Community Advisory Group (CAG)

• Collecting usage data (BR2) via SUSHI

• Identifying issues involved and resources required

Page 20: Summary

E-book trial

Page 21: Summary

Developing JUSP for different markets

• Responding to interest in JUSP

• Awareness of same issues being faced

• Expertise and enthusiasm of the JUSP team

Page 22: Summary

Pilot projects

Individual libraries– An overseas university

• University of Western Australia

– UK government libraries• Defence Science &

Technology Laboratory (DSTL)

Page 23: Summary

JUSPConsult

Page 24: Summary

Consultancy, advisory and support services

Justification• Advice on building the case

for a usage statistics portal, costs and efficiency gains

Implementation• Assistance with building a

platform, providing a hosted service or meeting individual requirements

Analysis and interpretation• Usage data analysis,

requirements gathering and presentation

Page 25: Summary

Next steps

• Continued development of pilot trials

• COUNTER Release 4

• Extending beyond JR1

• Aligning with wider projects and service

Page 26: Summary

Contact details

Web:jusp.mimas.ac.ukjuspconsult.mimas.ac.uk

Email:[email protected]@manchester.ac.uk

Page 27: Summary

Questions?