36

Click here to load reader

Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

  • Upload
    stephen

  • View
    213

  • Download
    1

Embed Size (px)

Citation preview

Page 1: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

This article was downloaded by: [Florida State University]On: 21 December 2014, At: 05:39Publisher: RoutledgeInforma Ltd Registered in England and Wales Registered Number: 1072954Registered office: Mortimer House, 37-41 Mortimer Street, London W1T 3JH,UK

Journal of Research onTechnology in EducationPublication details, including instructions forauthors and subscription information:http://www.tandfonline.com/loi/ujrt20

Developing and Validating aReliable TPACK Instrumentfor Secondary MathematicsPreservice TeachersJeremy Zelkowskia, Jim Gleasona, Dana C. Coxb &Stephen Bismarckc

a University of Alabamab Miami Universityc University of South Carolina—UpstatePublished online: 21 Feb 2014.

To cite this article: Jeremy Zelkowski, Jim Gleason, Dana C. Cox & StephenBismarck (2013) Developing and Validating a Reliable TPACK Instrument for SecondaryMathematics Preservice Teachers, Journal of Research on Technology in Education,46:2, 173-206, DOI: 10.1080/15391523.2013.10782618

To link to this article: http://dx.doi.org/10.1080/15391523.2013.10782618

PLEASE SCROLL DOWN FOR ARTICLE

Taylor & Francis makes every effort to ensure the accuracy of all theinformation (the “Content”) contained in the publications on our platform.However, Taylor & Francis, our agents, and our licensors make norepresentations or warranties whatsoever as to the accuracy, completeness,or suitability for any purpose of the Content. Any opinions and viewsexpressed in this publication are the opinions and views of the authors, andare not the views of or endorsed by Taylor & Francis. The accuracy of theContent should not be relied upon and should be independently verified withprimary sources of information. Taylor and Francis shall not be liable for any

Page 2: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

losses, actions, claims, proceedings, demands, costs, expenses, damages,and other liabilities whatsoever or howsoever caused arising directly orindirectly in connection with, in relation to or arising out of the use of theContent.

This article may be used for research, teaching, and private study purposes.Any substantial or systematic reproduction, redistribution, reselling, loan,sub-licensing, systematic supply, or distribution in any form to anyone isexpressly forbidden. Terms & Conditions of access and use can be found athttp://www.tandfonline.com/page/terms-and-conditions

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 3: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

Volume 46 Number 2 | Journal of Research on Technology in Education | 173

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

JRTE | Vol. 46, No. 2, pp. 173–206 | ©2013 ISTE | iste.org/jrte

Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

Jeremy ZelkowskiJim Gleason

University of Alabama

Dana C. CoxMiami University

Stephen BismarckUniversity of South Carolina—Upstate

Abstract

Within the realm of teaching middle and high school mathematics, the abil-ity to teach mathematics effectively using various forms of technology is now more important than ever, as expressed by the recent development of the Common Core State Standards for Mathematical Practice. This article pres-ents the development process and the results from 15 institutions and more than 300 surveys completed by secondary mathematics preservice teach-ers. The results suggest that technological, pedagogical, and content knowl-edge; technology knowledge; content knowledge; and pedagogical knowledge constructs are valid and reliable, whereas pedagogical content knowledge, technological content knowledge, and technological pedagogical knowledge domains remain difficult for preservice teachers to separate and self-report. (Keywords: TPACK, survey, secondary, mathematics education, teacher edu-cation, preservice)

With the recent development and adoption of the Common Core State Standards for Mathematics (CCSSM) (NGA & CCSSO, 2010) come new standards for mathematical practice that re-emphasize

the importance of technological tools in mathematics classrooms. This renewed commitment to helping students model their world and strategi-cally select tools for mathematical pursuits essentially requires teachers of mathematics to be knowledgeable and effective at teaching mathematics with technology and assessing the learning of mathematics by students who use technology when doing mathematics (Stein, Smith, Henningsen, & Silver 2010). The CCSSM are not the first set of standards that attend to the importance of integrating technology into the teaching and learning of mathematics. The National Council of Teachers of Mathematics (NCTM)

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 4: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

174 | Journal of Research on Technology in Education | Volume 46 Number 2

Zelkowski, Gleason, Cox, & Bismarck

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

and the Association of Mathematics Teacher Educators (AMTE) have also made specific recommendations regarding the role of technology in the mathematics classroom and for the development of mathematics teachers. NCTM’s Technology Principle provides a vision of mathematics teaching where technology is essential, driving curricular changes in the mathematics that can and should be learned in the classroom. The technology principle states, “Technology is essential in teaching and learning mathematics; it influences the mathematics that is taught and enhances students’ learning” (NCTM, 2000, p. 24). To enact this vision, AMTE (2006) recommends that “mathematics teacher preparation programs must ensure that all math-ematics teachers and teacher candidates have opportunities to acquire the knowledge and experiences needed to incorporate technology in the context of teaching and learning mathematics” (p.1). The specialized knowledge that teachers require to effectively integrate technology into teaching practices is currently referred to as technological, pedagogical, and content knowledge (TPACK) (Mishra & Koehler, 2006).

Given that some mathematics teachers utilize technology effectively and consistently in their classrooms, while others seem to avoid technology, we must ask whether teachers’ values and beliefs are at the heart of this paradox. Further, because of the growing national perspective that K–12 students should have access to technological tools and the hope that this access will greatly influence the way mathematics is taught and learned, it behooves current teacher preparation programs to examine their current programs and ensure that they are producing technologically effective mathematics teachers. The purpose of this study was to develop, establish the reliability of, and validate constructs for, an instrument that would enable preservice sec-ondary mathematics teacher educators to examine their preservice teachers’ (PSTs’) self-perceptions of TPACK and evaluate their program’s effectiveness at developing PSTs’ TPACK.

Theoretical Background

Defining Mathematical TPACK TPACK emerged in the literature as a theoretical framework for teachers’ knowledge of effective technology use in the classroom for teaching and learning (Mishra & Koehler, 2006). The TPACK framework was developed by expanding the construct of pedagogical content knowledge (PCK) to in-clude the integration of technology knowledge (TK) for teaching (see Figure 1). Shulman’s (1986) construct of PCK referred to the teaching knowledge that specifically blends both content knowledge (CK) and pedagogical knowledge (PK). Specific to mathematics, CK and PK, as well as PCK, frameworks have been proposed and studied relatively extensively during the past two decades (Ball, 1990; Ball, Hill, & Bass, 2002; Brown & Borko, 1992; Hill, 2011; Lampert & Ball, 1998; Ma, 1999; Silverman & Thompson,

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 5: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

Volume 46 Number 2 | Journal of Research on Technology in Education | 175

Developing and Validating a Reliable TPACK Instrument

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

2008; Strawhecker, 2005; Van Voorst, 2004), though more so at the primary grade levels than secondary grade levels.

Because of this work, teacher education programs have placed much greater attention on the development of these teacher knowledge do-mains (Ball, Lubienski, & Mewborn, 2001; Dede & Soybas, 2011; Fryk-holm & Glasson, 2005; Kahan, Cooper, and Bethea, 2003; Wilson, Floden, Ferrini-Mundy, 2001). With the inclusion of TK as an additional teacher knowledge domain and the advanced technology developments for teaching mathematics, teacher education programs for teachers of mathematics, now must examine their program effectiveness at develop-ing seven domains of knowledge in PSTs. These seven domains include, TK, PK, CK, PCK, technological pedagogical knowledge (TPK), techno-logical content knowledge (TCK), and TPACK (see Figure 1).

For a more general treatment and description of each of these domains, we refer the reader to Schmidt, Baran, Thompson, Mishra, Koehler, and Shin (2009) and the literature above. Here, we elaborate on TK, TCK, and TPACK in the context of secondary mathematics. To begin crafting these working definitions, we acknowledge the considerable amount of accomplished work in the field (see “TPACK in secondary mathematics” below).

TK in secondary mathematics. TK in secondary mathematics refers to knowledge of technologies that are relevant to secondary mathematics

Figure 1. Seven knowledge domains central to the TPACK framework (http://tpack.org).

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 6: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

176 | Journal of Research on Technology in Education | Volume 46 Number 2

Zelkowski, Gleason, Cox, & Bismarck

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

teaching and learning. The list of technologies specific to secondary math-ematics classrooms is ever growing and includes technologies developed specifically for the mathematics classroom as well as mainstream tech-nologies that serve computational and/or instructional needs. In the first category, the list surely includes computer algebra systems (CAS), dynamic mathematical software (e.g., Fathom, Geogebra, Geometer’s Sketchpad, TI-Nspire Teacher’s Edition Software), online Java applets, and graphing handhelds. In the second category, the list minimally includes calculat-ing devices, spreadsheets, data collection devices such as calculator-based laboratories (CBLs), interactive whiteboards, personal response systems, and many others.

TCK in secondary mathematics. TCK in secondary mathematics refers to the knowledge that incorporating technology fundamentally changes the mathematics that students can learn. Three examples will highlight the impact of technology on mathematical content in the classroom. First, dynamic mathematical software transforms the classroom into a mathemati-cal laboratory, giving students the opportunity to engage with mathematical concepts and relationships in an empirical environment. This, in turn, opens up new perspectives and problem-solving techniques that would otherwise not be possible. Second, graphing handheld devices give students immedi-ate access to multiple representations of mathematical relationships. These representations and the connections between them are not only reasoning tools, but also communication tools that assist students as they learn to negotiate shared meanings. Third, spreadsheets and data collection devices give students access to larger, more complex data sets with which to model real-world phenomena. These data sets would be incomprehensible to stu-dents working exclusively with pencil and paper.

TPACK in secondary mathematics. Beyond the questions of what technolo-gies are relevant to the teaching of mathematics and what mathematics can be taught with the inclusion of technology is the question of how best to leverage the power of technology in the classroom. Knowledge about how technologies can influence the teaching and learning of mathematics is referred to as TPACK. Teachers of mathematics must possess enough spe-cialized knowledge to make critical classroom decisions of pedagogy with respect to mathematical content and appropriate technologies. Internation-ally, research has shown students learn more mathematics and learn it more deeply by using technology effectively and appropriately during the teaching and learning of mathematics—including some studies showing the closing of achievement gaps in underrepresented students (Barkatsas, Kasimatic, & Gialamas, 2009; Bos, 2007; Doerr & Zangor, 2000; Duda, 2011; Dugdale & Kibbey, 1990; Dunham & Dick, 1994; Guckin & Morrison, 1991; Hol-lebrands, 2003; Kaput, 1995; Li & Ma, 2010; López, 2010; Mitchell, Bailey, & Monroe, 2007; Page, 2002; Schmidt, Kohler, & Moldenhauer, 2009; Weng-linsky & Educational Testing Service, 1998; Zbiek, 1998). However, some

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 7: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

Volume 46 Number 2 | Journal of Research on Technology in Education | 177

Developing and Validating a Reliable TPACK Instrument

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

studies have shown little to no achievement effects specific to the integration of technology into the mathematics classroom (Kelly, Carnine, Gersten, & Grossen, 1986; Shapley, Sheehan, Maloney, & Caranikas-Walker, 2011; Torff & Tirotta, 2010). The distinguishing factor between the research findings appears to be fundamentally grounded in how the technology is used peda-gogically by the teacher, their beliefs about technology, and whether students are using the device to make mathematical connections and check for un-derstanding rather than simply for computation, also referred to as answer-getting. Teachers’ effective technology implementation in the mathematics classroom has been slow, according to the literature, for various reasons, including but not limited to: (a) resistance to change, (b) lack of confidence, (c) access, (d) usefulness, (e) fears (i.e., technical failures, students know-ing more about technology than the teacher), and (f) inhibitors (Beaudin & Bowers, 1997; Keating & Evans, 2001; Leatham, 2007; Li, 2003; McKinney & Frazier, 2008; Norton, McRobbie, & Cooper 2000; Pierce & Ball, 2009; Quinn, 1998; Stallings, 1995; Thomas, 2006; Thompson, 1992; Ursini, San-tos, & Juarez Lopez, 2005).

There are many examples in the field of studies that purport to document, define or teach the appropriate use of technology in the mathematics class-room (i.e., Demana & Waits 1990, 1998; Drier, 2001; Kutzler, 2000; Niess, 2001; 2005; Salinger, 1994; Shoaf, 2000). With this existing knowledge of research, the AMTE (2009) adopted the Mathematics TPACK framework for teacher educators. Specifically, the Mathematics TPACK framework consists of four essential components: (1) Design and develop technology-enhanced mathematics learning environments and experiences, (2) Facilitate math-ematics instruction with technology as an integrated tool, (3) Assess and evaluate technology-enriched mathematics teaching and learning, and (4) Engage in ongoing professional development to enhance TPACK. Further, Niess and her colleagues (2009) produced the TPACK development model for teachers. They indicate that:

… many mathematics teachers’ PCK lacks a solid and consistent inte-gration of modern digital technologies in mathematics curriculum and instruction. Technologies, such as dynamic geometry tools or advanced graphing calculators with computer algebra systems (CAS), are primarily used for modeling and providing examples, where students imitate the actions and use the technologies for verification, demonstration, and drill and practice. In essence then, while digital technologies have evolved, strategies for their effective integration into the learning of mathematics have not evolved as rapidly (p. 6).

With the recent development and adoption of the CCSSM, along with the advancement of educational technologies specific to mathematics, the expectation to teach and learn mathematics with the appropriate use of technology has increased dramatically, as can be seen in the standards for

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 8: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

178 | Journal of Research on Technology in Education | Volume 46 Number 2

Zelkowski, Gleason, Cox, & Bismarck

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

mathematical practices (NGA & CCSSO, 2010). It has become increasingly important for researchers to develop instruments for monitoring the profes-sional growth of preservice mathematics teachers’ ability and self-efficacy to use technology effectively and appropriately (or not) for teaching mathemat-ics (i.e., TPACK). As mathematics teachers increase their self-awareness and confidence in the use of technology in the classroom, many increase their pedagogical toolbox that utilizes technology to enhance the teaching and learning of mathematics (Lawless & Pellegrino, 2007). Since its incep-tion in 2005, TPACK has been of great interest among those concerned with preservice preparation programs, yet much of the large-scale studies have been along the lines of Schmidt et al. (2009) and involve preservice elemen-tary generalists, where preservice populations are large, rather than second-ary specialists in mathematics with preservice populations that are generally small at universities across the United States. What was not clear in the literature is a systematic way to longitudinally study TPACK development on a grand scale. The purpose of our study was to create an instrument for such research.

Developing an Instrument to Measure TPACK for Preservice Secondary Mathematics TeachersIn their recent review of the TPACK literature, Young, Young, and Shaker (2012) studied the instruments and methods of studying TPACK in pre-service teachers. Their conclusion revealed no specialized instrument for secondary mathematics TPACK. The TPACK development model (Neiss et al., 2009) has been a framework for TPACK development in preservice teachers, yet the research community lacks the instruments to measure and monitor TPACK development. Some performance assessments exist for use with preservice elementary (i.e., Angeli & Valandies, 2009) or inservice (i.e., Bos, 2011; Harris, Grandgenett, & Hofer, 2010) teachers, and these generally take the form of surveys.

One benefit of using a survey format is that it measures self-efficacy. The self-efficacy of preservice and inservice teachers influences whether or not they can and will implement technology supported classroom lessons (Bowers & Stephens, 2011; Dunham & Henessey, 2008; Kastberg & Leatham, 2005; Kendal & Stacey, 2001; Harris & Hofer, 2011; Niess, 2013; Tharp, Fitzsimmons, & Ayers, 1997; Zbiek & Hollebrands, 2008). Özgün-Koca, Meagher, and Edwards (2010) used the mathematics technology attitudes survey (MTAS) as part of their secondary mathematics preservice prepara-tion program. Hofer and Grandgenett (2012) conducted a one-year longi-tudinal study on a cohort of alternative certification preservice teachers but limited the study to a small sample size (<10) and less than one year of data collection. Surveys have been designed and tested in large-scale studies on a number of different populations, including K–12 online educators (Archam-bault & Crippen, 2009; Archambault & Barnett, 2010) as well as preservice

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 9: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

Volume 46 Number 2 | Journal of Research on Technology in Education | 179

Developing and Validating a Reliable TPACK Instrument

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

teachers, including both primary and secondary in Singapore (Koh, Chai, & Tsai, 2010) and the United States (Sahin, 2011) and U.S. elementary PSTs (Lux, Bangert, & Whittier, 2011; Schmidt et al., 2009).

Although we can assume that Koh, Chai, and Tsai (2010) and Sahin (2011) included secondary preservice mathematics teachers in their samples, neither distinguished this group within the whole. There is an absence in the literature with respect to a valid and reliable survey to monitor and assess TPACK development in secondary mathematics PSTs. In a recent review of the literature on measuring TPACK among PSTs, Abbitt (2011) conjectures that the survey developed by Schmidt et al. (2009) “may also provide a solid foundation for similar surveys that would be applicable to PSTs who will be teaching other content areas and grade levels,” (p. 292), a conjecture that is echoed by Niess (2011), who adds that more work is needed to validate such an instrument, including more large-scale studies.

Our current large-scale study aimed to establish a valid and reliable instrument for monitoring and assessing TPACK development in secondary mathematics PSTs, including the seven interrelated knowledge domains that TPACK encompasses. In theory, a valid and reliable TPACK instrument for this population would permit secondary mathematics teacher preparation programs to longitudinally study program effectiveness at TPACK develop-ment, fine-tune program components, conduct large-scale multi-institution-al research studies, and make changes to maximize TPACK development in next generation of teachers of mathematics.

MethodOur goal was to develop an instrument that answers the need to monitor the development of TPACK in preservice secondary mathematics teachers throughout preparation programs. Adapting the TPACK survey for elemen-tary PSTs (Schmidt et al., 2009), this project focused on developing and validating a reliable content-specific survey for the population of interest: preservice secondary mathematics teachers.

Instrument DevelopmentWe began by examining the reliable and nonreliable items from the Schmidt et al. (2009) survey by deleting items that did not specifically pertain to the teaching of secondary mathematics. Specifically, we deleted the items relat-ing to social studies, literacy, and science, as the instrument was established for generalists at the elementary level. Next, we wrote an additional 22 items to fill gaps in the seven knowledge domain constructs (TK, PK, CK, PCK, TPK, TCK, TPACK) by focusing on specific content areas in mathematics. We wrote these items to adhere to item development guidelines (Fink, 2003; Fink & Kosecoff, 1998; Fowler, 2008; Pattern, 2001). Six researchers, includ-ing two external to the project, with expertise in secondary mathematics education reviewed all of the items to be included in the survey for content

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 10: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

180 | Journal of Research on Technology in Education | Volume 46 Number 2

Zelkowski, Gleason, Cox, & Bismarck

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

validity (Lawshe, 1975), even though the Schmidt survey development team had previously vetted many of the survey items. We then revised seven items based on the feedback from the expert panel. Three items received minor editing, such as moving a word to a different location in the item or using a word with similar meaning that was more specific to mathematics. Six items received major editing, such as completely restructuring the item wording. After this revision, the final survey instrument contained 62 items consisting of 8 TK items, 8 CK items, 8 PK items, 7 PCK items, 7 TPK items, 12 TCK items, and 12 TPACK items. Each item used a 5-point Likert scale response (SD = “strongly disagree,” D = “disagree,” N = “neither agree nor disagree,” A = “agree,” SA = “strongly agree”).

For the purposes of this study, we designed a cover sheet to collect addi-tional demographic data, including participants’ age, level in college, practi-cum experience, gender, and ethnicity. We also designed an additional cover sheet to gather information about the context in which data was collected, including the name of the course for which the survey was administered, date of survey, whether students completed a technology course, if technol-ogy is used in mathematics content courses, and the grade-band of teacher certification for secondary mathematics at each respective institution.

ParticipantsThe research team sought to collect data from a variety of secondary math-ematics teacher preparation programs across the United States in an effort to maximize the diversity in our national sample. For this project, diversity included size of institution, type of institution, size of secondary mathemat-ics education program, demographics of student population, experience of faculty teaching program courses, and geographic location. We contacted 24 secondary mathematics education faculty around the country to inquire whether they would be willing to administer the survey to their PSTs. Half (12) agreed to organize administering the survey in addition to the three institutions of the co-authors. The sample was a mix of convenience and strategically chosen institutions to produce a diverse sample. To maximize authentic responses and completion rates, paper surveys were administered in classes rather than online out-of-class surveys (Adams & Gale, 1982; Lefever, Dal, & Matthiasdottir, 2007; Norris & Conn, 2005). Faculty were encouraged to arrange administering the survey at the start of class but were permitted to administer the survey at their convenience. Faculty had gradu-ate students or teaching assistants administer the survey in accordance with IRB requirements. Due to these efforts, our survey completion return rate was over 90% and as we will mention later, the surveys returned seemed to contain predominantly legitimate responses.

We collected 315 surveys completed by PSTs prior to student teaching from 15 institutions across the United States (see Table 1) with class sizes ranging from 4 to 26 students during the 2010–11 academic year. Survey

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 11: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

Volume 46 Number 2 | Journal of Research on Technology in Education | 181

Developing and Validating a Reliable TPACK Instrument

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

administrators at different sites reported that the survey took an average of 15–20 minutes for participants to complete after the instructions were read and any questions were addressed. Upon receiving the completed surveys from each institution, the project’s principal investigator worked with his graduate assistant to begin coding the paper survey responses into spread-sheet format. We created a simple spreadsheet for all 315 returned surveys, which included the demographic responses in Table 1 as well as the re-sponses to the 62-item survey. To guarantee accurate coding, an additional graduate student randomly selected 25% of the surveys and rechecked the original spreadsheet entries. Four entry cells were miscoded (<0.001% error rate), so the graduate student rechecked the remaining 75% of the surveys,

Table 1. Demographic Data of Total Responses

% by region % by institution type

Variable % MW NE SE W RI RT

Age

< 19 0.3 1.5 0.0 0.0 0.0 0.7 0.0

19–22 68.3 89.7 66.7 62.8 40.6 67.1 68.2

23–26 19.0 5.9 15.8 22.6 43.8 19.9 20.3

27–30 2.9 0.0 3.5 3.6 6.3 2.1 4.1

> 30 9.5 2.9 12.3 11.7 9.4 11.0 8.8

Class

Sophomore 8.6 36.8 3.5 0.0 0.0 6.8 12.2

Junior 17.8 16.2 26.3 16.8 0.0 17.1 16.2

Senior 50.8 41.2 40.4 61.3 50.0 43.8 58.8

Graduate 20.6 2.9 26.3 21.2 50.0 30.8 12.2

Gender

Male 32.4 39.7 24.6 32.1 43.8 32.2 25.0

Female 67.3 60.3 75.4 67.9 53.1 67.8 75.0

Practicum

Yes 58.4 46.0 44.0 70.0 59.0 69.9 47.3

No 41.6 54.0 56.0 30.0 41.0 30.1 52.7

Ethnicity

African Am 4.1 36.8 1.8 5.8 3.1 6.2 2.0

Am Indian 0.3 0.0 0.0 0.7 0.0 0.0 0.7

Asian 1.3 0.0 3.5 0.7 3.1 2.1 0.7

Hisp/Latino 3.5 1.5 1.8 0.7 25.0 0.7 6.1

Pac Islander 1.0 0.0 1.8 0.7 3.1 0.7 0.7

White 85.7 94.1 84.2 88.3 56.3 87.0 83.1

Other 2.2 1.5 5.3 0.7 3.1 0.7 6.1

Notes:. Regional classifications (Midwest, Northeast, Southeast, West) based on higher education consortia’s of states (e.g. West-ern Interstate Commission on Higher Education). Institutional classification based on Carnegie classification and labeled as either RI=Research Intensive, RT=Some research and/or Teaching only. Some percentages may not sum to 100 due to non-responses and/or rounding.

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 12: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

182 | Journal of Research on Technology in Education | Volume 46 Number 2

Zelkowski, Gleason, Cox, & Bismarck

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

found an additional eight cell-entry errors, and corrected them. The project PI then randomly selected 10% of the surveys and found no entry errors. This verification method of checking, correcting, and rechecking authenti-cates that the 19,530 cells in the spreadsheet had virtually a zero error rate for entry before the analyses.

The final sample reported in Table 1 reflects 49.6% of responses from sec-ondary mathematics PSTs at eight research-intensive, institutions with the remaining amount coming from seven institutions with some research activ-ity and/or teaching only. The national sample consisted of responses from institutions in the Midwest (23.1%), Northeast (19.4%), Southeast (46.6%), and West (10.9%) regions of the United States. Our sample oversampled approximately 10% higher in the Southeast and 10% lower in the West when examining the 2010 U.S. census populations. The Midwest and Northeast are within 2% of census populations to the region.

Determining the Final Sample for AnalysisAs the participants were asked to complete the survey in class, they may have felt compelled or obligated to complete the survey but lacked the motivation to engage in the survey. To account for this possible lack of engagement, we used four criteria for removing cases from the sample that demonstrated lack of engagement. First, the research team determined that students who gave nearly the same response for all items on the survey dem-onstrated a lack of engagement, so we examined all 315 completed surveys for less than 10% variance in the responses. Four surveys met the first filter criteria with respective total survey variances of responses with 1.6%, 3.2%, 4.9%, and 6.4% (the next closest variation was 10.8%). We deleted these four surveys. Similarly, if we found zero variance on any one of three pages of survey items (approximately 20 items), we removed those surveys. We removed 14 remaining surveys that met this criterion, 4 of which had more than one page with zero variance, indicating nonengagement in reading the items and distinguishing any variation on roughly 20 items. Third, we per-formed a manual inspection of surveys to look for patterning to determine if students appeared to fill in their responses in unlikely ways. We determined patterning on two surveys not already deleted. One survey had responses that were 1, 2, 3, 4, 5, 4, 3, 2, 1, etc., item to item, and the other alternated between two responses for every item. Fourth, we applied the criterion of completeness (all 62 items answered) and removed one incomplete survey from the sample. In all, we removed 21 surveys (6.67% of the sample) from the 315 surveys completed, resulting in a sample of 294 surveys for the main analysis on 62 items (18,228 cells).

Data AnalysisTo determine the structure of the TPACK instrument, we randomly split the data into two equal groups based on the overall TPACK summation score of

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 13: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

Volume 46 Number 2 | Journal of Research on Technology in Education | 183

Developing and Validating a Reliable TPACK Instrument

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

Likert scale items so that the distribution between the two groups was the same. We used an exploratory factor analysis (EFA) with the first group of 147 student surveys to determine which items should be retained for the fi-nal instrument. We used a confirmatory factor analysis on the second group of 147 student surveys to verify the structure obtained through the explor-atory factor analysis. This method gives more strength and validity to the research method and results (Fink, 2003; Litwin, 1995; Thompson, 2004). To determine the internal reliability of each of the subscales, we used a graded response model (Samejima, 1996) with the full sample of 294 surveys to analyze the instrument in addition to the traditional Cronbach alpha for reliability measures.

Results

Exploratory Factor Analysis (EFA)Based on the theoretical construct that TPACK includes TK, CK, PK, PCK, TCK, and TPK, our initial analysis explored the possibility of seven fac-tors, one for each trait. However, the initial EFA on the data showed distinct factors only for TK (Factor 2), CK (Factor 5), PK (Factor 3), and TPACK (Factor 1) (see Table 2, p. 184). The items written to measure PCK, TCK, and TPK loaded on various factors with no obvious pattern or more than three items loading on a single factor. Therefore, we decided to remove the PCK, TCK, and TPK items from the remainder of the analysis and to leave the measurement of the two-knowledge latent traits for future studies with a new item development process.

After the removal of the PCK, TCK, and TPK items from the analysis, we completed an exploratory factor analysis on the remaining items with four factors. The four-factor analysis matches the data well, with TK loading on Factor 2, CK loading on Factor 3, PK loading on Factor 4, and TPACK load-ing on Factor 1 (see Table 3, p. 186).

Confirmatory Factor Analysis (CFA) and Internal Reliability MeasuresTo finalize the instrument, we performed a CFA using a four-factor structure and included all items with a factor loading greater than or equal to 0.500. During the analysis, we monitored particularly a number of fit indices and sta-tistics. Thompson (2004) indicates the CFA fit indices measures and statistics of particular interest for our particular instrument type are the chi-squared statistical significance test, non-normed fit index (NNFI), comparative fit in-dex (CFI), and root-mean-square error of approximation (RMSEA). Further, Thompson indicates that all four should be carefully examined and considered to determine a final model. The non-normed and comparative fit statistics should be close to 0.95 or greater to indicate an excellent model fit (0.90–0.95 is acceptable to good). The RMSEA estimates how well the model parameters will perform at reproducing the population covariances. The RMSEA should

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 14: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

184 | Journal of Research on Technology in Education | Volume 46 Number 2

Zelkowski, Gleason, Cox, & Bismarck

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

Table 2. Factor Loadings for EFA with Varimax Rotation of Seven Factors

Item Factor 1 Factor 2 Factor 3 Factor 4 Factor 5 Factor 6 Factor 7

TK01 -0.175 0.738 -0.226

TK02 -0.274 0.837

TK03 0.860 -0.187 0.141

TK04 0.845 -0.121

TK05 0.695

TK06 0.739 -0.134 0.169 0.103

TK07 0.228 0.513 -0.102 0.157

TK08 0.120 -0.421 0.203 -0.113

CK09 -0.196 -0.102 0.732

CK10 0.103 0.899 -0.276

CK11 0.778 -0.265

CK12 0.503 0.170

CK13 -0.159 0.564 0.199

CK14 0.128 0.491 0.345

CK15 -0.183 -0.127 0.257 0.640

CK16 -0.154 0.119 0.451 0.203

PK17 0.627 0.130 -0.101

PK18 -0.120 0.828

PK19 -0.168 0.147 0.907 -0.179

PK20 -0.149 0.766

PK21 0.252 0.613 -0.165 -0.125

PK22 0.443 0.262

PK23 0.436 -0.144 0.191

PK24 0.630 0.132 0.110

PCK25 0.386 0.146 0.233

PCK26 0.168 0.659

PCK27 -0.180 0.110 0.751

PCK28 0.342 0.235 0.209 -0.167

PCK29 0.193 0.129 0.189 0.151 0.109

PCK30 -0.151 0.180 0.591

PCK31 -0.161 -0.173 0.130 0.501 0.144 0.291 0.152

TCK32 0.318 -0.100 0.219 -0.162 0.441

TCK33 -0.145 0.536 -0.194

TCK34 0.494 -0.107 0.375 -0.123 -0.323

TCK35 0.608 0.124 -0.259 0.102 -0.390

TCK36 0.872 -0.116 -0.128

TCK37 -0.111 0.754

TCK38 0.349 0.148 -0.171 -0.213

TPK39 0.733 -0.111 -0.121 -0.206

TPK40 0.713 -0.127 0.146 0.150 -0.128 -0.172

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 15: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

Volume 46 Number 2 | Journal of Research on Technology in Education | 185

Developing and Validating a Reliable TPACK Instrument

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

be zero if we can produce the exact population covariance. Values approxi-mately 0.06 or less are considered very reasonable for a model fit. Once we determined the finalized instrument through a CFA, we checked internal reliability for measuring the internal consistency on each unidimensional construct. The statistics for internal reliability are the Cronbach’s alpha, stan-dard error of measurement, and number of respondents within one and two standard deviations (Cronbach & Shavelson, 2004; Green, Lissitz, & Mulaik, 1977; Schmitt, 1996). Cronbach’s alpha is used to indicate the degree to which set of items measures the unidimensional construct and is acceptable when greater than 0.70, good when greater than 0.80, and excellent when greater than 0.90. Reporting the standard errors and respondents within one and two standard deviations demonstrates a normalized set of responses from the population and determines if there is a strong skew in the respons-es for each construct.

Confirmatory factor analysis (CFA). Using the second group of 147 students, the CFA was conducted. The initial analysis had some minor questions with fit to explore and refine. Using an iterative process of removing items with

Table 2 continued

Item Factor 1 Factor 2 Factor 3 Factor 4 Factor 5 Factor 6 Factor 7

TPK41 0.488 0.131 0.192 -0.195

TPK42 0.485 0.198 0.106 0.109 -0.303

TPK43 0.583 0.156 -0.236

TPK44 -0.134 0.256

TPK45 0.207 0.542 0.113

TPK46 0.143 0.188 -0.112 0.120 0.322

TPK47 0.457 0.146 0.129 0.159 -0.172

TPK48 0.124 -0.187

TPK49 0.254 -0.118 0.262 -0.105

TPK50 0.147 0.159 -0.112 -0.254

TPACK51 0.759 -0.115 0.149 0.122

TPACK52 0.987 -0.150 -0.197 -0.156

TPACK53 1.009 -0.141 -0.120 -0.147 -0.101

TPACK54 0.495 0.255 -0.119

TPACK55 0.717 0.149 0.219

TPACK56 0.505 0.204 0.354

TPACK57 0.321 0.157 0.184 -0.193 0.534 0.335

TPACK58 0.210 -0.131 0.748 0.269

TPACK59 0.758 0.249

TPACK60 0.623 -0.185 0.119 0.219 0.259

TPACK61 0.119 0.913 0.491

TPACK62 -0.121 0.802 0.390

Notes. Factor loadings ≥ 0.500 are in boldface. Test of the hypothesis is that seven factors are sufficient. Chi-square statistic is 2375.54 on 1,478 degrees of freedom, p<0.001.

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 16: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

186 | Journal of Research on Technology in Education | Volume 46 Number 2

Zelkowski, Gleason, Cox, & Bismarck

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

Table 3. Factor Loadings for EFA with Varimax Rotation of Four Factors.

Item Factor 1 Factor 2 Factor 3 Factor 4

TK01 -0.163 0.706

TK02 -0.248 0.848 0.102

TK03 0.843 -0.174

TK04 0.828 -0.114

TK05 0.115 0.669

TK06 0.713

TK07 0.212 0.462

TK08 0.120 -0.405 0.161

CK09 -0.197 0.755

CK10 0.725

CK11 -0.113 0.655

CK12 0.536 0.108

CK13 0.636

CK14 0.118 0.662

CK15 -0.113 -0.118 0.794

CK16 -0.166 0.632

PK17 0.667

PK18 0.801

PK19 -0.118 0.158 0.822

PK20 -0.123 0.758

PK21 0.195 -0.176 0.656

PK22 0.172 0.166 0.479

PK23 0.479

PK24 0.104 0.571

TPACK51 0.864 -0.103 -0.120

TPACK52 0.857 -0.151

TPACK53 0.923 -0.143

TPACK54 0.516 0.233 -0.154

TPACK55 0.798

TPACK56 0.239 0.490

TPACK57 0.570 0.154

TPACK58 0.386 0.236

TPACK59 0.829

TPACK60 0.750 0.133 -0.147

TPACK61 0.507 0.216 -0.141

TPACK62 0.281 0.100 0.305

Notes. Factor loadings ≥ 0.500 are in boldface. Test of the hypothesis is that four factors are sufficient. Chi-square statistic is 861.66 on 492 degrees of freedom, p<0.001.

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 17: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

Volume 46 Number 2 | Journal of Research on Technology in Education | 187

Developing and Validating a Reliable TPACK Instrument

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

small loadings onto the desired factor and/or loadings onto undesired fac-tors, we established a final instrument with 22 items (6 TK, 5 CK, 5 PK, 6 TPACK). The iterative process included determining items that loaded onto multiple factors. For example, we deleted TPACK62 due to cross-loading almost equally on two constructs. Items with less than a 0.5 factor loading were considered for deletion. Remaining items with a factor loading over 0.5 were examined for high correlations with other items. The remaining items deleted were due to high correlations with other times. For example, TPACK54 had a high correlation (>0.8) with TPACK51. We determined that TPACK51 was a much stronger and more meaningful item than TPACK54. Further, some of the indices indicated that the fit would be greatly improved by removing CK10 and allowing TPACK59 and TPACK60 to interact. This is natural, as the geometry and algebra items for TPACK should have some correlation for secondary mathematics PSTs because much of the content is interconnected (NCTM, 2000). The CFA for four factors is presented in Table 4 (p. 188).

Further, as the analysis of the instrument includes using a graded re-sponse model to estimate the standard error curve, each of the subscales must be confirmed to be unidimensional. To verify the unidimensionality, we analyzed each subscale (TK, CK, PK, TPACK) using all 294 students.

Internal reliability. The Technology Knowledge (TK) subscale had a Cronbach’s alpha of 0.8899 with the graded response model (see Figure 2, p. 189) giving a standard error of measurement below 0.35 for most of the respondents. For the TK subscale, 74% of the respondents were within one standard deviation of the mean, and 96% were within two standard devia-tions of the mean.

The Content Knowledge (CK) subscale had a Cronbach’s alpha of 0.8554 with the graded response model (see Figure 3, p. 189) giving a standard error of mea-surement below 0.53 for participants within one standard deviation of the mean. For the CK subscale, 73% of the respondents were within one standard deviation of the mean, and 99% were within two standard deviations of the mean.

The Pedagogical Knowledge (PK) subscale had a Cronbach’s alpha of 0.8768 with a graded response model (see Figure 4, p. 190), giving a stan-dard error of measurement below 0.5 for all of the participants. For the PK sub-scale, 72% of the respondents were within one standard deviation of the mean and 93% were within two standard deviations of the mean.

The TPACK subscale had a Cronbach’s alpha of 0.8966 with a graded re-sponse model (see Figure 5, p. 190) giving a standard error of measurement below 0.5 for almost all of the participants. For the TPACK subscale, 71% of the respondents were within one standard deviation of the mean, and 94% were within two standard deviations of the mean.

As Cronbach’s alpha for each of the four subscales is above a 0.85 for each construct, the internal reliability is very good to excellent in measure for differences at the group level, and therefore the subscales meet the desired

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 18: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

188 | Journal of Research on Technology in Education | Volume 46 Number 2

Zelkowski, Gleason, Cox, & Bismarck

requirements for survey research instruments (DeVellis, 2011). The iterative item removal process resulted in a model with good to excellent fit indices. The model fit index statistics are presented in Table 5 (p. 191), and the over-all model is presented as a diagram (see Figure 6, p. 192).

The chi-square statistic for the PK subscale is worth further discussion. Thompson (2004) indicates this statistic is important but not the end-all statistic. However, all the fit indices need serious consideration as well. Our

Table 4. Factor Loadings for CFA with Varimax Rotation of Four Factors after Iterative Process for Item Removal

Item TPACK TK CK PK

TK01 0.663

TK02 -0.179 0.801

TK03 0.853 -0.161

TK04 0.831 -0.132

TK05 0.133 0.665

TK06 0.667

CK09 -0.228 0.774

CK10 0.826

CK11 0.754

CK12 0.579 0.102

CK13 0.593

CK14 0.108 0.555

PK17 0.142 0.625

PK18 0.821

PK19 -0.109 0.133 0.827

PK20 -0.121 0.730

PK21 0.208 0.608

TPACK51 0.801 -0.122

TPACK52 0.880 -0.103

TPACK53 0.929

TPACK55 0.710 0.111

TPACK59 0.750

TPACK60 0.678 0.139 0.147 -0.152

Factor 1 Factor 2 Factor 3 Factor 4

SS loadings 3.949 3.541 2.963 2.738

Proportion Variance 0.172 0.154 0.129 0.119

Cumulative Variance 0.172 0.326 0.454 0.573

Factor Correlations

Factor 1 -0.333 0.438 0.347

Factor 2 -0.187 -0.348

Factor 3 0.348

Notes. Factor loadings ≥ 0.500 are in boldface. Test of the hypothesis is that four factors are sufficient. Chi-square statistic is 272.38 on 167 degrees of freedom, p<0.001.

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 19: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

Volume 46 Number 2 | Journal of Research on Technology in Education | 189

Developing and Validating a Reliable TPACK Instrument

Figure 2. TK subscale standard error and participant distribution.

Figure 3. CK subscale standard error and participant distribution.

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 20: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

190 | Journal of Research on Technology in Education | Volume 46 Number 2

Zelkowski, Gleason, Cox, & Bismarck

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

Figure 4. PK subscale standard error and participant distribution.

Figure 5. TPACK subscale standard error and participant distribution.

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 21: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

Volume 46 Number 2 | Journal of Research on Technology in Education | 191

Developing and Validating a Reliable TPACK Instrument

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

analysis leads us to conclude that the PK subscale is likely valid because all three fit indices are the strongest compared to the TK, CK, and TPACK subscales. The NFI, CFI, and RMSEA are all beyond excellent with respect to acceptable ranges for these fit indices for the PK subscale. Table 6 (p. 193) presents the final model with parameter estimates for the finalized items on the TPACK instrument for the TK, CK, PK, and TPACK constructs. The data presented includes the parameter estimates, standard errors, z-values, and model error term (e#) on each item.

Discussion

Key Findings and Extending the TPACK LiteratureThe methodology and data analysis provide a consistent protocol for de-veloping and testing an instrument that can be used for examining TPACK knowledge in secondary PSTs. In doing so, we report three major findings. First and foremost, the results from this research indicate a proficient instru-ment for monitoring and measuring secondary mathematics PSTs’ self-assessment of TK, PK, CK, and TPACK. This is the first survey instrument designed specifically for secondary mathematics PSTs. Unlike prior studies (e.g. Archambault & Crippen, 2009; Schmidt et al., 2009) that focused on teachers of multiple disciplines, our study was unique because it focused specifically on the next generation of teachers of mathematics.

Second, our findings revealed that PCK, TPK, and TPK items did not produce a measurable factor for each knowledge domain. This finding brings up an issue with trying to develop an instrument that can identify knowledge domains that overlap between two categories for secondary mathematics teachers. This result adds a finding to the literature that differs from the instrument of Schmidt et al. (2009), in which elementary PSTs reported a reliable seven-factor instrument. TPK, PCK, and TCK remain elusive to quantify and measure well through the survey items we tested. In fact, TPK, PCK, and TCK self-reporting may be more difficult for preservice mathematics teachers to grasp during preservice preparation. TPK, PCK, and TCK require PSTs to mentally remove one knowledge construct, which is abstract and likely difficult for PSTs to do when responding to these items. This could stem from the compartmentalized nature of many programs that

Table 5. Confirmatory Factor Analysis Statistical Fit Indices for Final Model and Subscales

Description Chi-square df p NNFI CFI RMSEA

Four-factor model 374.40 202 0.00000 0.940 0.947 0.052

TK subscale 37.67 9 0.00002 0.947 0.968 0.104

CK subscale 18.82 5 0.00207 0.945 0.972 0.097

PK subscale 8.60 5 0.12607 0.990 0.995 0.049

TPACK subscale 44.55 8 0.00000 0.932 0.964 0.125

Notes. Four-factor model statistics and each subscale unidimensionality statistics.

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 22: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

192 | Journal of Research on Technology in Education | Volume 46 Number 2

Zelkowski, Gleason, Cox, & Bismarck

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

Figu

re 6

. Con

firm

ator

y fa

ctor

ana

lysi

s m

odel

dia

gram

for

final

four

-fac

tor

inst

rum

ent.

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 23: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

Volume 46 Number 2 | Journal of Research on Technology in Education | 193

Developing and Validating a Reliable TPACK Instrument

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

divide pedagogy, technology, and content knowledge into separate courses. However, this may be supported by the reliability of the TK, PK, and CK items. Additional reasons may include the fact that specialized secondary education content areas are so content focused that PSTs may have difficulty removing one of the three knowledge domains in their own minds.

Lastly, the correlation (see Figure 6) between the four factors revealed that TK and PK have a very low correlation (0.020), whereas the others all have moderate correlations (≥0.348). So we examined these factor correlations further. We interpret these correlational findings to imply that TPACK is cor-related moderately with TK, PK, and CK. The relationship is nearly the same for each factor. This indicates that TPACK development may be influenced by each factor about the same. CK’s correlation to PK and TK is about the same, implying that CK may be influenced by both PK and TK about the same. However, the low TK–PK relationship may indicate that preservice secondary mathematics teachers (PSMTs) who have made up their minds about pedago-gy strategies (PK) may not be strong candidates for believing they do not need

Table 6. Final Survey Instrument Model Statistics for Four-Factor Model

ItemParameterCoefficient

StandardError (SE)

Error Terme#

Error Term SE

TK01 0.629 0.049 0.472 0.042

TK02 0.638 0.040 0.251 0.026

TK03 0.744 0.046 0.316 0.033

TK04 0.765 0.049 0.395 0.039

TK05 0.679 0.049 0.442 0.041

TK06 0.575 0.039 0.273 0.026

CK09 0.312 0.029 0.174 0.016

CK11 0.411 0.030 0.145 0.016

CK12 0.451 0.032 0.161 0.018

CK13 0.360 0.031 0.184 0.018

CK14 0.524 0.044 0.364 0.035

PK17 0.488 0.034 0.205 0.020

PK18 0.500 0.032 0.156 0.017

PK19 0.499 0.034 0.195 0.019

PK20 0.505 0.032 0.163 0.017

PK21 0.461 0.034 0.220 0.021

TPACK51 0.465 0.034 0.230 0.021

TPACK52 0.568 0.032 0.126 0.015

TPACK53 0.557 0.033 0.156 0.017

TPACK55 0.553 0.036 0.214 0.022

TPACK59 0.531 0.038 0.260 0.024

TPACK60 0.515 0.041 0.342 0.031

Notes. Z-values for all 22 items >8.97 with p<0.0001. The error term (SE) values for the interaction between TPACK59 and TPACK60 is 0.113 (0.021).

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 24: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

194 | Journal of Research on Technology in Education | Volume 46 Number 2

Zelkowski, Gleason, Cox, & Bismarck

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

TK. On the other hand, if PSMTs have strong TK, it does not imply that they see they have strong pedagogy beliefs. The key finding of all this sense-making is that the fact that CK is correlated well with PK and TK may simply imply that higher CK is likely when PSMTs have higher TK and PK, instead of that a PSMT with lower TK or PK will have lower content knowledge. It could be possible that PSMTs with more education methods courses may have stronger CK, as K–12 content is focused on more in those courses, as opposed to math-ematics department upper-division content courses (e.g., analysis, abstract algebra). PSMTs who have completed upper-division mathematics courses may perceive themselves as having lower CK because their strong self-efficacy in mathematics may be challenged in theoretical-based courses.

With these results, our goal of developing a valid and reliable instrument to monitor and assess preservice mathematics teachers’ TPACK develop-ment within teacher education programs was lucrative for a four-factor in-strument. Although no instrument is perfect, our instrument provides a ro-bust survey aimed at one specific population that traditionally is very small at respective institutions of higher education, making it nearly impossible to largely quantify TPACK components through research at one or even a few institutions. This adds to the TPACK literature with a large-scale quantita-tive study of a historically low population, institution to institution, in the United States. Our instrument builds on the work of Mishra and Koehler (2005a, 2005b, 2006) through the adapting of the Schmidt et al. work (2009). We extend the TPACK literature domain with this instrument as compared to previous study findings for different populations.

The main goal of this research project was to develop a valid and reliable instrument for the population of secondary mathematics PSTs to help math-ematics teacher educators (MTEs) monitor and assess their own programs’ ability to develop TPACK in their PSTs. The data from this research resulted in survey items to reliably monitor and assess the constructs of TK, PK, CK, and TPACK so historically low populations of students at institutions can use a valid and reliable instrument to examine their program effectiveness. This instrument can be used to better inform a single program course and can provide overall information for an entire program. If the instrument is applied as a pre–post survey for a program, the information does not provide detailed information regarding course specific material. A limitation arises with programs where secondary mathematics PSMTs’ content knowl-edge is targeted by a mathematics department and their pedagogical knowl-edge is targeted by an education department, and there is a lack of commu-nication between both departments. This is not uncommon in secondary mathematics teacher preparation programs across the United States.

This study was not specifically about what students know, but rather an objective assessment of what they think they know. PSMTs’ personal beliefs change positively and negatively over time during preservice prepara-tion. This instrument gives researchers and educators the ability to reliably

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 25: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

Volume 46 Number 2 | Journal of Research on Technology in Education | 195

Developing and Validating a Reliable TPACK Instrument

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

measure these beliefs regarding TPACK and the contributing factors of PK, TK, and CK during preservice preparation programs. PSMTs can easily be overconfident or lack confidence. Our experiences provide us with knowl-edge that PSMTs can be exposed to new ways of thinking during coursework and improve or diminish their sense of self-efficacy. Learning can go in both directions, so our study provides an opportunity for programs and educators to understand their learning environment to improve TPACK development in PSMTs.

LimitationsA limitation worth noting is the potential of variation in the administration of surveys. Specific instructions were given to each site, and we rely on the professionals following those guidelines, and we must rely on an assumption that the surveys were administered in a similar fashion at each institution. More significant, we were unable to isolate PCK, TCK, and TPK knowledge domains specifically in the population of secondary mathematics PSTs. We acknowledge that this is limiting in terms of the use of our survey, yet we see unearthing such a limitation as a key finding in expanding the TPACK literature domain.

Implications for Practice, Directions for Future ResearchIt is imperative that we develop TPACK in our preservice mathematics teachers in order to fully enact the vision for teaching set forth by AMTE, CCSSM, and NCTM. The finalized instrument presented in Table 6 (p. 193) represents a valid, reliable, and manageable 22-item survey (or 16 items without TPACK items) with which secondary mathematics teacher educa-tion programs can monitor and assess PSMTs’ perceptions of their devel-opment of TPACK. Further, for the successful integration of technology into the teaching of secondary mathematics, preservice teachers need to be comfortable with their own perceived abilities to do so during their prepara-tion programs. Our instrument will give researchers and MTEs the ability to conduct TPACK research in preservice courses and programs using a variety of methods.

First, using the instrument in a pre–post format can allow for and detect changes over a semester course or during the student teaching internship. We see the instrument, for example, aiding in discovering which cooperat-ing teachers are effective at helping grow TPACK in PSMTs during their student teaching. This instrument can also be used longitudinally over two years. This will provide insights into when and what aspects of preparation programs improve PSMTs’ perceptions of TPACK knowledge growth. We also see this instrument as useful in the development of TPACK practices in the classroom accompanied by an observational instrument (e.g., Hofer et al., 2011). This research needs to be done in the future. Additionally, there may be potential in using the instrument to help PSTs recognize their own

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 26: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

196 | Journal of Research on Technology in Education | Volume 46 Number 2

Zelkowski, Gleason, Cox, & Bismarck

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

professional growth and development. Helping PSTs overcome their precon-ceived beliefs is difficult, yet this instrument can aid in the development and assessment early in a preparation program.

In the future, mixed-methods research designs would be strong for exam-ining TPACK in PSTs using this survey along with classroom observational techniques, including interviews and video analyses. Moreover, combining our instrument with observations during student teaching may provide a rich data set of both quantitative and qualitative data. Such a rich set of data will allow for strong, valid, and more generalizable research findings. Fur-ther, it would be useful to begin to index specific curricular events (e.g., in-structional strategies, assignments, goals) or clinical experiences for growth in TPACK and improvement in the technological self-efficacy of PSTs. This can be accomplished by using this instrument as an element of large-scale comparative studies.

There is a great deal of work to be done with respect to TPACK develop-ment in the next generation of secondary mathematics preservice teachers. We hope our instrument will aid in some of that work.

AcknowledgmentsThe authors wish to thank individuals who helped us by administering or arranging the admin-istration of our survey to the secondary mathematics preservice teachers at their respective insti-tutions. Thank you, Ginny Bohme, Daniel Brahier, Anna Marie Conner, Kelly Edenfield, Carla Gerberry, Shannon Guerrero, Gary Martin, Ginger Rhodes, Wendy Sanchez, Tommy Smith, Toni Smith, Jami Stone, Anthony Thompson, and Jan Yow. We also wish to express our gratitude to the preservice secondary mathematics teachers who engaged and diligently completed surveys at each of the respective institutions of higher education. Finally, we wish to thank Barbara and Robert Reys for their visionary mission to link early career mathematics teacher educators through the STaR program to advance the future of the mathematics education field. Without their hard work and the mentoring professionals of the STaR program, the data collection for this project may have taken years and the professional relationships might never have been established.

Author Notes

Jeremy Zelkowski, PhD, is an assistant professor of secondary mathematics education in the Department of Curriculum and Instructions in the College of Education at the University of Alabama, Tuscaloosa. His research interests focus on TPACK development in preservice teachers, appropriate use of technology in teaching mathematics, mathematical college readiness, and pro-fessional development learning communities. Please address correspondence regarding this article to Dr. Jeremy Zelkowski, Department of Curriculum and Instruction, University of Alabama, 902 University Blvd, Tuscaloosa, AL 35487-0232. Email: [email protected]

Jim Gleason, PhD, is an associate professor in the Department of Mathematics in the College of Arts & Sciences at the University of Alabama, Tuscaloosa. His research interests focus on teacher content knowledge, assessment, and psychometrics pertaining to mathematics teaching and learning.

Dana C. Cox, PhD, is an assistant professor in the Department of Mathematics in the College of Arts & Sciences at Miami University, Oxford, Ohio. Her research interests focus on curriculum design, TPACK development in preservice teachers, and studying the emergent professional vision of teaching with technology held by preservice teachers.

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 27: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

Volume 46 Number 2 | Journal of Research on Technology in Education | 197

Developing and Validating a Reliable TPACK Instrument

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

Stephen Bismarck, PhD, is an assistant professor of middle/secondary mathematics education in the School of Education at the University of South Carolina—Upstate in Spartanburg. His research interests focus on appropriate use of technology, developing mathematical knowledge for teaching with preservice teachers, and exploring best practices for building conceptual knowledge.

ReferencesAbbitt, J. T. (2011). Measuring technological pedagogical content knowledge in preservice

teacher education: A review of current methods and instruments. Journal of Research on Technology in Education, 43(4), 281–300.

Adams, L. L. M., & Gale, D. (1982). Solving the quandary between questionnaire length and response rate in education research. Research in Higher Education, 17(3), 231–240.

Angeli, C., & Valanides, N. (2009). Epistemological and methodological issues for the conceptualization, development, and assessment of ICT-TPCK: Advances in technological pedagogical content knowledge (TPCK). Computers & Education, 52(1), 154–168

Archambault, L. M., & Barnett, J. H. (2010). Revisiting technological pedagogical content knowledge: Exploring the TPACK framework. Computers & Education, 55(4), 1656–1662.

Archambault, L., & Crippen, K. (2009). Examining TPACK among K–12 online distance educators in the United States. Contemporary Issues in Technology and Teacher Education (CITE Journal), 9(1), 71–88.

Association of Mathematics Teacher Educators (AMTE). (2006). Preparing teachers to use technology to enhance the learning of mathematics. Retrieved August 1, 2008, from http://www.amte.net/resources

Association of Mathematics Teacher Educators (AMTE). (2009). Mathematics TPACK framework. Retrieved October 13, 2009, from http://www.amte.net/resources

Ball, D. L. (1990) Prospective elementary and secondary teachers’ understanding of division. Journal for Research in Mathematics Education, 21(2), 132–144.

Ball, D. L., Hill, H. C., & Bass, H. (2002). Developing measures of mathematics knowledge for teaching. Ann Arbor, MI: Study of Instructional Improvement.

Ball, D. L., Lubienski, S., & Mewborn, D. (2001). Research on teaching mathematics: The unsolved problem of teachers’ mathematical knowledge. In V. Richardson (Ed.), Handbook of research on teaching (4th ed.). New York: Macmillan.

Barkatsas, A., Kasimatis, K., & Gialamas, V. (2009). Learning secondary mathematics with technology: Exploring the complex interrelationship between students’ attitudes, engagement, gender and achievement. Computers & Education, 52(3), 562–570.

Beaudin, M., & Bowers, D. (1997). Logistics for facilitating CAS instruction. In J. Berry, J. Monaghan, M. Kronfellner, & B. Kutzler (Eds.), The state of computer algebra in mathematics education (pp. 126–135). Lancashire, England: Chartwell-York.

Bos, B. (2007). The effect of the Texas Instrument interactive instructional environment on the mathematical achievement of eleventh grade low achieving students. Journal of Educational Computing Research, 37(4), 351–368.

Bos, B. (2011). Professional development for elementary teachers using TPACK. Contemporary Issues in Technology and Teacher Education, 11(2), 167–183.

Bowers, J., & Stephens, B. (2011). Using technology to explore mathematical relationships: A framework for orienting mathematics courses for prospective teachers. Journal of Mathematics Teacher Education, 14(4), 285–304. doi:10.1007/s10857-011-9168-x

Brown, C. A., & Borko, H. (1992). Becoming a mathematics teacher. In D. A. Grouws (Ed.), Handbook of research on mathematics teaching and learning. New York: Macmillan.

Cronbach, L. J., & Shavelson, R. J. (2004). My current thoughts on coefficient alpha and successor procedures. Educational and Psychological Measurement, 64(3), 391–418.

Dede, Y., & Soybas, D. (2011). Preservice mathematics teachers’ experiences about function and equation concepts. EURASIA Journal of Mathematics, Science & Technology Education, 7(2), 89–102.

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 28: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

198 | Journal of Research on Technology in Education | Volume 46 Number 2

Zelkowski, Gleason, Cox, & Bismarck

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

Demana, F., & Waits, B. K. (1990). Implementing the standards: The role of technology in teaching mathematics. Mathematics Teacher, 83(1), 27–31.

Demana, F., & Waits, B. K. (1998). The role of graphing calculators in mathematics reform. Retrieved from ERIC database. (ED458108).

DeVellis, R. F. (2011). Scaled development: Theory and applications. Thousand Oaks, CA: Sage Publications.

Doerr, H. M., & Zangor, R. (2000). Creating meaning for and with the graphing calculator. Educational Studies in Mathematics, 41(2), 143–163.

Drier, H. S. (2001). Teaching and learning mathematics with interactive spreadsheets. School Science and Mathematics 101(4), 170–179.

Duda, J. (2011). Mathematical creative activity and the graphic calculator. International Journal for Technology in Mathematics Education, 18(1), 3–14.

Dugdale, S., & Kibbey, D. (1990). Beyond the evident content goals—Part I: Tapping the depth and flow of the educational undercurrent. Journal of Mathematical Behavior, 9, 201–228.

Dunham, P. H., & Dick, T. P. (1994). Connecting research to teaching: Research on graphing calculators. Mathematics Teacher, 87(6), 440–445

Dunham, P., & Hennessey, S. (2008). Equity and the use of educational technology in mathematics. In M. K. Heid & G. W. Blume (Eds.), Research on technology and the teaching and learning of mathematics: Vol. 1. Research syntheses (pp. 345–418). Charlotte, NC: Information Age.

Fink, A. (2003). How to manage, analyze, and interpret survey data (2nd edition). Thousand Oaks, CA: Sage Publications.

Fink, A., & Kosecoff, J. (1998). How to conduct surveys: A step-by-step guide. Thousand Oaks, CA: Sage Publications.

Fowler, F. J. (2008). Survey research methods (4th edition). Thousand Oaks, CA: Sage Publications.

Frykholm, J., & Glasson, G. (2005). Connecting science and mathematics instruction: Pedagogical context knowledge for teachers. School Science and Mathematics, 105(3), 127–140.

Green, S. B., Lissitz, R. W., & Mulaik, S. A. (1977). Limitations of coefficient alpha as an index of test unidimensionality. Educational and Psychological Measurement, 37(4), 827–838.

Guckin, A., & Morrison, D. (1991). Math*Logo: A project to develop proportional reasoning in college freshmen. School Science and Mathematics, 91(2), 77–81.

Harris, J. B., & Hofer, M. J. (2011). Technological pedagogical content knowledge (TPACK) in action: A descriptive study of secondary teachers’ curriculum-based, technology-related instructional planning. Journal of Research on Technology in Education, 43(3), 211–229.

Harris, J. B., Grandgenett, N., & Hofer, M. (2010, March). Testing a TPACK-based technology integration assessment rubric. Paper presented at the annual meeting of the Society for Information Technology and Teacher Education (SITE), San Diego, CA.

Hill, H. C. (2011). The nature and effects of middle school mathematics teacher learning experiences. Teachers College Record, 113(1), 205–234.

Hofer, M. & Grandgenett, N. (2012). TPACK development in teacher education: A longitudinal study of preservice teachers in a secondary M.A.Ed. program. Journal of Research on Technology in Education 45(1), 83–106.

Hofer, M., Grandgenett, N., Harris, J., & Swan, K. (2011). Testing a TPACK-based technology integration observation instrument. In C. D. Maddux, D. Gibson, B. Dodge, M. Koehler, P. Mishra, & C. Owens (Eds.). Research highlights in technology and teacher education 2011 (pp. 39–46). Chesapeake, VA: Society for Information Technology & Teacher Education.

Hollebrands, K. F. (2003). High school students’ understandings of geometric transformations in the context of a technological environment. Journal of Mathematical Behavior, 22(1), 55–72.

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 29: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

Volume 46 Number 2 | Journal of Research on Technology in Education | 199

Developing and Validating a Reliable TPACK Instrument

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

Kahan, J. A., Cooper, D. A., & Bethea, K. A. (2003). The role of mathematics teachers’ content knowledge in their teaching: A framework for research applied to a study of student teachers. Journal of Mathematics Teacher Education, 6(3), 223–252. doi: 10.1023/A:1025175812582.

Kaput, J. (1995). Creating cybernetic and psychological ramps for the concrete to the abstract: Examples from multiplicative structures. In D. Perkins, J. Schwartz, M West, & M. Wiske (Eds.), Software goes to school: Teaching for understanding with new technologies (pp. 130–154). New York: Oxford University Press.

Kastberg, S., & Leatham, K. (2005). Research on graphing calculators at the secondary level: Implications for mathematics teacher education. Contemporary Issues in Technology and Teacher Education [Online serial], 5(1).

Keating, T., & Evans, E. (2001). Three computers in the back of the classroom: Preservice teachers’ conceptions of technology integration. In R. Carlsen, N. Davis, J. Price, R. Weber, & D. Willis (Eds.), Society for Information Technology and Teacher Education Annual, 2001 (pp. 1671–1676). Norfolk, VA: Association for the Advancement of Computing in Education.

Kelly, B., Carnine, D., Gersten, R., & Grossen, B. (1986). The effectiveness of videodisc instruction in teaching fractions to learning-disabled and remedial high school students. Journal of Special Education Technology, 7(2), 5–17.

Kendal, M. & Stacey, K. (2001). The impact of teacher privileging on learning differentiation with technology. International Journal of Computers for Mathematical Learning, 6(2), 143–165.

Koehler, M. J., & Mishra, P. (2005a). Teachers learning technology by design. Journal of Computing In Teacher Education, 21(3), 94–102.

Koehler, M. J., & Mishra, P. (2005b). What happens when teachers design educational technology? The development of technological pedagogical content knowledge. Journal of Educational Computing Research, 32(2), 131–152.

Koh, J. L., Chai, C. S., & Tsai, C. C. (2010). Examining the technological pedagogical content knowledge of Singapore pre-service teachers with a large-scale survey. Journal of Computer Assisted Learning, 26(6), 563–573. doi:10.1111/j.1365-2729.2010.00372.x

Kutzler, B. (2000). The algebraic calculator as a pedagogical tool for teaching mathematics. International Journal of Computer Algebra in Mathematics Education, 7(1), 5–23.

Lampert, M., & Ball, D. L. (1998). Teaching, multimedia, and mathematics: Investigations of real practice. New York: Teacher’s College Press.

Lawless, K. A., & Pellegrino, J. W. (2007). Professional development in integrating technology into teaching and learning: Knowns, unknowns, and ways to pursue better questions and answers. Review of Educational Research, 77(4), 575–614.

Lawshe, C. H. (1975). A quantitative approach to content validity. Personnel Psychology, 28(4), 563–575.

Leatham, K. R. (2007). Pre-service secondary mathematics teachers’ beliefs about the nature of technology in the classroom. Canadian Journal of Science, Mathematics, and Technology Education. 7(2/3), 183–207.

Lefever, S., Dal, M., & Matthiasdottir, A. (2007). Online data collection in academic research: Advantages and limitations. British Journal of Educational Technology, 38(4), 574–582.

Li, Q. (2003). Would we teach without technology? A professor's experience of teaching mathematics education incorporating the internet. Educational Research, 45(1), 61–77.

Li, Q., & Ma, X. (2010). A meta-analysis of the effects of computer technology on school students’ mathematics learning. Educational Psychology Review, 22(3), 215–243.

Litwin, M. S. (1995). How to measure survey reliability and validity. Thousand Oaks, CA: Sage Publications.

López, O. S. (2010). The digital learning classroom: Improving English language learners’ academic success in mathematics and reading using interactive whiteboard technology. Computers & Education, 54(4), 901–915. doi:10.1016/j.compedu.2009.09.019

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 30: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

200 | Journal of Research on Technology in Education | Volume 46 Number 2

Zelkowski, Gleason, Cox, & Bismarck

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

Lux, N., Bangert, A., & Whittier, D. (2011). The development of an instrument to assess preservice teachers’ technological pedagogical content knowledge. Journal of Educational Computing Research, 45(4), 415–431.

Ma, L. P. (1999). Knowing and teaching elementary mathematics. Mahwah, NJ: Lawrence Erlbaum.

McKinney, S., & Frazier, W. (2008). Embracing the principles and standards for school mathematics: An inquiry into the pedagogical and instructional practices of mathematics teachers in high-poverty middle schools. Clearing House: A Journal of Educational Strategies, Issues, and Ideas. 81(5), 201–210.

Mishra, P., & Koehler, M. J. (2006). Technological pedagogical content knowledge: A framework for integrating technology in teachers’ knowledge. Teachers College Record, 108(6), 1017–1054.

Mitchell, B., Bailey, J. L., & Monroe, E. (2007). Integrating technology and a standards-based pedagogy in a geometry classroom: A mature teacher deals with the reality of multiple demands and paradigm shifts. Computers in the Schools, 24(1–2), 75–91.

National Council of Teachers of Mathematics. (2000). Principles and standards for school mathematics. Reston, VA: Author.

National Governors Association (NGA) Center for Best Practices & Council of Chief State School Officers (CCSSO). (2010). The Common Core State Standards initiative: Preparing America’s students for college and career. Washington, DC: Author.

Niess, M. L. (2001). A model for integrating technology in preservice science and mathematics content-specific teacher preparation. School Science & Mathematics, 101(2), 102–109.

Niess, M. L. (2005). Preparing teachers to teach science and mathematics with technology: Developing a technology pedagogical content knowledge. Teaching and Teacher Education, 21(5), 509–523.

Niess, M. L. (2011). Investigating TPACK: Knowledge growth in teaching with technology. Journal for Educational Computing Research. 44(3), 299–317.

Niess, M. (2013). Central component descriptors for levels of technological pedagogical content knowledge. Journal of Educational Computing Research, 48(2), 173–198.

Niess, M. L., Ronau, R. N., Shafer, K. G., Driskell, S. O., Harper S. R., Johnston, C., Browning, C., Özgün-Koca, S. A., & Kersaint, G. (2009). Mathematics teacher TPACK standards and development model. Contemporary Issues in Technology and Teacher Education [Online serial], 9(1), 4–24.

Norris, J., & Conn, C. (2005). Investigating strategies for increasing student response rates to online-delivered course evaluations. Quarterly Review of Distance Education, 6(1), 13–29.

Norton, S., McRobbie, C. J., & Cooper, T. J. (2000). Exploring secondary math teachers’ reasons for not using computers in their teaching: Five case studies. Journal of Research on Computing in Education. 33(1), 87–109.

Özgün-Koca, S. A., Meagher, M., & Edwards, M. T. (2010). Preservice teachers’ emerging TPACK in a technology-rich methods class. Mathematics Educator, 19(2), 10–20.

Page, M. S. (2002). Technology-enriched classrooms: Effects on students of low socioeconomic status. Journal of Research on Technology in Education, 34(4), 389–409.

Pattern, M. L. (2001). Questionnaire research: A practical guide (2nd edition). Los Angeles: Pyrczak Publishing.

Pierce, R., & Ball, L. (2009). Perceptions that may affect teachers’ intention to use technology in secondary mathematics classes. Educational Studies in Mathematics, 71(3), 299–317.

Quinn, R. J. (1998). Technology: PSTs’ beliefs and the influence of a mathematics methods course. Clearing House. 71(6), 375–377.

Sahin, I. (2011). Development of survey of technological pedagogical and content knowledge (TPACK). Turkish Online Journal of Educational Technology—TOJET, 10(1), 97–105.

Salinger, G. L. (1994). Educational reform movements and technology education. Technology Teacher, 53(5), 6–8

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 31: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

Volume 46 Number 2 | Journal of Research on Technology in Education | 201

Developing and Validating a Reliable TPACK Instrument

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

Samejima, F. (1996). Graded response model. In W. J. van der Linden & R. K. Hambleton (Eds.), Handbook of modern item response theory. New York: Springer-Verlag.

Schmidt, D. A., Baran, E., Thompson, A. D., Mishra, P., Koehler, M. J., & Shin, T. S. (2009). Technological pedagogical content knowledge (TPACK): The development and validation of an assessment instrument for preservice teachers. Journal of Research on Technology in Education, 42(2), 123–149.

Schmidt, K., Kohler, A., & Moldenhauer, W. (2009). Introducing a computer algebra system in mathematics education—Empirical evidence from Germany. International Journal for Technology in Mathematics Education, 16(1), 11–26

Schmitt, N. (1996). Uses and abuses of coefficient alpha. Psychological Assessment, 8(4), 350–353.

Shapley, K., Sheehan, D., Maloney, C., & Caranikas-Walker, F. (2011). Effects of technology immersion on middle school students’ learning opportunities and achievement. Journal of Educational Research, 104(5), 299–315. doi:10.1080/00220671003767615

Shoaf, M. (2000). A capstone course for pre-service secondary mathematics teachers. International Journal of Mathematical Education in Science & Technology, 31(1), 151–160.

Shulman, L. S. (1986). Those who understand: Knowledge growth in teaching. Educational Researcher, 15(2), 4–14.

Silverman, J., & Thompson, P. W. (2008). Toward a framework for the development of mathematical knowledge for teaching. Journal of Mathematics Teacher Education, 11(6), 499–511. DOI: 10.1007/s10857-008-9089-5.

Stallings, L. L. (1995). Teachers’ stories learning to use computing technologies in mathematics teaching. Unpublished doctoral dissertation, University of Georgia, Athens.

Stein, M. K., Smith, M. S., Henningsen, M. A., & Silver, E. A. (2010) Implementing standards-based mathematics instruction: A casebook for professional development. New York: Teachers College Press.

Strawhecker, J. (2005). Preparing elementary teachers to teach mathematics: How field experiences impact pedagogical content knowledge. Issues in the Undergraduate Mathematics Preparation of School Teachers, 4(Curriculum), 12 pp. Retrieved from ERIC database. (EJ835513).

Tharp, M. L., Fitzsimmons, J. A., & Ayers, R. L. B. (1997). Negotiating a technological shift: Teacher perception of the implementation of graphing calculators. Journal of Computers in Mathematics and Science Teaching, 16(4), 551–575.

Thomas, M. O. J. (2006). Teachers using computer in mathematics: A longitudinal study. In C. Hoyles, J. Lagrance, L. H. Son & Sinclair (Eds.), Proceedings for the 17th ICMI Study Conference: Technology Revisited, Hanoi University of Technology, December 3–8, 2006 (c17).

Thompson, A. G. (1992). Teachers’ beliefs and conceptions: A synthesis of the research. In D.A. Grouws (Ed.), Handbook of Research in Mathematics Teaching and Learning (pp. 127–146). New York: MacMillan.

Thompson, B. (2004). Exploratory and confirmatory factor analysis: Understanding concepts and applications. Washington, DC: American Psychological Association.

Torff, B., & Tirotta, R. (2010). Interactive whiteboards produce small gains in elementary students’ self-reported motivation in mathematics. Computers & Education, 54(2), 379–383. doi:10.1016/j.compedu.2009.08.019

Ursini, S., Santos, D., & Juarez Lopez, J. A. (2005). Teachers’ resistance using technology: Source of ideas for a pedagogical proposal. In F. Olivero & R. Sutherland (Eds.), Proceedings of the 7th International Conference on Technology in Mathematics Teaching (ICTMT) (pp. 189–197). Bristol, UK: University of Bristol.

Van Voorst, C. (2004). Capstone mathematics courses for teachers. Issues in the Undergraduate Mathematics Preparation of School Teachers, 4(Curriculum), 11 pp. Retrieved from ERIC database. (EJ835511).

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 32: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

202 | Journal of Research on Technology in Education | Volume 46 Number 2

Zelkowski, Gleason, Cox, & Bismarck

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

Wenglinsky, H., & Educational Testing Service. (1998). Does it compute? The relationship between educational technology and student achievement in mathematics. Retrieved from ERIC database. (ED425191).

Wilson, S., Floden, R., & Ferrini-Mundy, J. (2001). Teacher preparation research: Current knowledge, gaps and recommendations. A research report prepared for the U.S. Department of Education by the Center for the Study of Teaching and Policy in collaboration with Michigan State University. Retrieved from http://depts.washington.edu/ctpmail/PDFs/TeacherPrep-WFFM-02-2001.pdf

Young, J. R., Young, J. L., & Shaker, Z. (2012). Describing the preservice teacher technological pedagogical content knowledge (TPACK) literature using confidence intervals. Tech Trends, 56(5), 25–33.

Zbiek, R. (1998). Prospective teachers’ use of computing tools to develop and validate functions as mathematical models. Journal for Research in Mathematics Education, 29(2), 184–201.

Zbiek, R. M., & Hollebrands, K. (2008). A research-informed view of the process of incorporating mathematics technology into classroom practice by inservice and preservice teachers. In M. K. Heid & G. W. Blume (Eds.), Research on technology and the teaching and learning of mathematics: Vol. 1. Research syntheses (pp. 287–344). Charlotte, NC: Information Age.

Manuscript received April 3, 2013 | Initial decision April 26, 2013 | Revised manuscript accepted May 20, 2013

Appendix A

Administered SurveyThank you for taking time to complete this survey. Please answer each question to the best of your knowledge. You should answer demographic information first, then read each item and choose your first belief. You need not spend any lengthy time on any one item. You should be finished in about 15 minutes.

Your thoughtfulness and candid responses will be greatly appreciated. Your confi-dentiality will not be compromised and your name will not, at any time, be associated with your responses.

Your responses will be kept completely confidential and will not influence your course grade.

Demographic Information

Age range:

a. Under 19b. 19–22c 23–26d. 27–30e. 30+

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 33: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

Volume 46 Number 2 | Journal of Research on Technology in Education | 203

Developing and Validating a Reliable TPACK Instrument

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

Year in college:

a. Freshmanb Sophomorec. Juniord. Seniore. Graduate student

Have you completed a practicum or clinical placement in a middle grades or high school math classroom?

a. Yesb. No

Gender:

a. Maleb. Femalec. I prefer not to say

Ethnicity:

a. African American or blackb. Alaskan Nativec. American Indiand. Asiane. Hispanic or Latinof. Pacific Islanderg. White or Caucasianh. Other: _____________i. I prefer not to say

Technology is a broad concept that can mean a lot of different things. For the purpose of this questionnaire, technology is referring to digital technology/technologies—that is, the digital tools we use, such as computers, laptops, iPods, handhelds, interactive whiteboards, computer software programs, graphing calculators, etc. Please answer all of the questions, and if you are uncertain of or neutral about your response, you may always select “Neither agree nor disagree.”

All items appeared as a SD, D, N, A, SA response set up.

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 34: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

204 | Journal of Research on Technology in Education | Volume 46 Number 2

Zelkowski, Gleason, Cox, & Bismarck

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

TK1 I know how to solve my own technical problems.

TK2 I can learn technology easily.

TK3 I keep up with important new technologies.

TK4 I frequently play around with the technology.

TK5 I know about a lot of different technologies.

TK6 I have the technical skills I need to use technology.

TK7 I have had sufficient opportunities to work with different technologies.

TK8 When I encounter a problem using technology, I seek outside help.

CK9 I have sufficient knowledge about mathematics.

CK10 I can use mathematical ways of thinking.

CK11 I have various strategies for developing my understanding of mathematics.

CK12 I know about various examples of how mathematics applies in the real world.

CK13 I have a deep and wide understanding of algebra.

CK14 I have a deep and wide understanding of geometry.

CK15 I have a deep and wide understanding of calculus.

CK16 I have a deep and wide understanding of advanced undergraduate mathematics.

PK17 I know how to assess student performance in a classroom.

PK18 I can adapt my teaching based upon what students currently understand or do not understand.

PK19 I can adapt my teaching style to different learners.

PK20 I can assess student learning in multiple ways.

PK21 I can use a wide range of teaching approaches in a classroom setting.

PK22 I am familiar with common student understandings and misconceptions.

PK23 I know how to organize and maintain classroom management.

PK24 I know when it is appropriate to use a variety of teaching approaches (e.g., problem/proj-ect-based learning, inquiry learning, collaborative learning, direct instruction) in a classroom setting.

PCK25 I know how to select effective teaching approaches to guide student thinking and learning in mathematics.

PCK26 I know different teaching approaches to teach ratio and proportion concepts.

PCK27 I know different strategies/approaches for teaching probability and statistics concepts.

PCK28 I know different strategies/approaches for teaching algebra concepts.

PCK29 I know different strategies/approaches for teaching geometry concepts.

PCK30 I know different strategies/approaches for teaching trigonometry concepts.

PCK31 I know different strategies/approaches for teaching calculus concepts.

TCK32 I know about technologies that I can use for understanding and doing ratio and proportion.

TCK33 I know about technologies that I can use for understanding and doing probability and statistics.

TCK34 I know about technologies that I can use for understanding and doing algebra.

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 35: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

Volume 46 Number 2 | Journal of Research on Technology in Education | 205

Developing and Validating a Reliable TPACK Instrument

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

TCK35 I know about technologies that I can use for understanding and doing geometry.

TCK36 I know about technologies that I can use for understanding and doing trigonometry.

TCK37 I know about technologies that I can use for understanding and doing calculus.

TCK38 I know that using appropriate technology can improve one’s understanding of mathematics concepts.

TPK39 I can choose technologies that enhance the teaching of a lesson.

TPK40 I can choose technologies that enhance students’ learning for a lesson.

TPK41 My teacher education program has caused me to think more deeply about how technology could influence the teaching approaches I use in my classroom.

TPK42 I am thinking critically about how to use technology in my classroom.

TPK43 I can adapt the use of the technologies that I am learning about to different teaching activities.

TPK44 Different teaching approaches require different technologies.

TPK45 I have the technical skills I need to use technology appropriately in teaching.

TPK46 I have the classroom management skills I need to use technology appropriately in teaching.

TPK47 I know how to use technology in different instructional approaches.

TPK48 My teaching approaches change when I use technologies in a classroom.

TPK49 Knowing how to use a certain technology means that I can use it for teaching.

TPK50 Different technologies require different teaching approaches.

TPACK51 I can use strategies that combine mathematics, technologies, and teaching approaches that I learned about in my coursework in my classroom.

TPACK52 I can choose technologies that enhance the mathematics for a lesson.

TPACK53 I can select technologies to use in my classroom that enhance what I teach, how I teach, and what students learn.

TPACK54 I can provide leadership in helping others to coordinate the use of mathematics, technologies, and teaching approaches at my school and/or district.

TPACK55 I can teach lessons that appropriately combine mathematics, technologies, and teaching approaches.

TPACK56 Integrating technology in teaching mathematics will be easy and straightforward for me.

TPACK57 I can teach lessons that appropriately combine ratio and proportion, technologies, and teaching approaches.

TPACK58 I can teach lessons that appropriately combine probability and statistics, technologies, and teaching approaches.

TPACK59 I can teach lessons that appropriately combine algebra, technologies, and teaching approaches.

TPACK60 I can teach lessons that appropriately combine geometry, technologies, and teaching approaches.

TPACK61 I can teach lessons that appropriately combine trigonometry, technologies, and teaching approaches.

TPACK62 I can teach lessons that appropriately combine calculus, technologies, and teaching approaches.

This survey was spread out in a four page format so items did not appear to run together.

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014

Page 36: Developing and Validating a Reliable TPACK Instrument for Secondary Mathematics Preservice Teachers

206 | Journal of Research on Technology in Education | Volume 46 Number 2

Zelkowski, Gleason, Cox, & Bismarck

Copyright © 2013, ISTE (International Society for Technology in Education), 800.336.5191(U.S. & Canada) or 541.302.3777 (Int’l), [email protected], iste.org. All rights reserved.

Appendix B

Finalized Survey ItemsTK1 I know how to solve my own technical problems.

TK2 I can learn technology easily.

TK3 I keep up with important new technologies.

TK4 I frequently play around with the technology.

TK5 I know about a lot of different technologies.

TK6 I have the technical skills I need to use technology.

CK9 I have sufficient knowledge about mathematics.

CK11 I have various strategies for developing my understanding of mathematics.

CK12 I know about various examples of how mathematics applies in the real world.

CK13 I have a deep and wide understanding of algebra.

CK14 I have a deep and wide understanding of geometry.

PK17 I know how to assess student performance in a classroom.

PK18 I can adapt my teaching based upon what students currently understand or do not understand.

PK19 I can adapt my teaching style to different learners.

PK20 I can assess student learning in multiple ways.

PK21 I can use a wide range of teaching approaches in a classroom setting.

TPACK51 I can use strategies that combine mathematics, technologies, and teaching approaches that I learned about in my coursework in my classroom.

TPACK52 I can choose technologies that enhance the mathematics for a lesson.

TPACK53 I can select technologies to use in my classroom that enhance what I teach, how I teach, and what students learn.

TPACK55 I can teach lessons that appropriately combine mathematics, technologies, and teaching approaches.

TPACK59 I can teach lessons that appropriately combine algebra, technologies, and teaching approaches.

TPACK60 I can teach lessons that appropriately combine geometry, technologies, and teaching approaches.

Dow

nloa

ded

by [

Flor

ida

Stat

e U

nive

rsity

] at

05:

39 2

1 D

ecem

ber

2014