227
1 THE ACQUISITION OF SCIENCE VOCABULARY: A MORPHOLOGICAL APPROACH TO CONTENT AREA LITERACY A thesis presented by Lisha Cabral to The School of Education In partial fulfillment of the requirements for the degree of Doctor of Education In the field of Education College of Professional Studies Northeastern University Boston, Massachusetts March 16, 2015

The acquisition of science vocabulary: a morphological ...rx914j94z/Fulltext.pdfautomaticity of understanding the morphology of words and improving their overall comprehension of text

  • Upload
    others

  • View
    4

  • Download
    0

Embed Size (px)

Citation preview

       

   

1  

THE ACQUISITION OF SCIENCE VOCABULARY: A MORPHOLOGICAL APPROACH TO CONTENT AREA LITERACY

A thesis presented by

Lisha Cabral

to The School of Education

In partial fulfillment of the requirements for the degree of Doctor of Education

In the field of

Education

College of Professional Studies Northeastern University Boston, Massachusetts

March 16, 2015

       

   

2  

Table of Contents

Table of Contents .......................................................................................................................... 2  List of Figures ................................................................................... Error! Bookmark not defined.  

Dedication ...................................................................................................................................... 4  Abstract .......................................................................................................................................... 5  

Chapter I: Introduction ................................................................................................................ 6  Statement of the Problem  ...............................................................................................................  6  Justification for the Research Problem  ..........................................................................................  9  Deficiencies in the Evidence  ..........................................................................................................  10  Relating the Discussion to Audiences  ...........................................................................................  11  Significance of the Research Problem: Implications of Poor Comprehension  .............................  12  Research Questions  .......................................................................................................................  16  Theoretical Framework  ................................................................................................................  16  

Chapter II: Literature Review ................................................................................................... 20  Scarborough’s Reading Rope Model of Reading  ..........................................................................  20  Conceptual nature of domain specific science vocabulary  ...........................................................  22  Extent of necessary vocabulary  ....................................................................................................  23  Need for strategies  ........................................................................................................................  24  

Chapter III: Research Design .................................................................................................... 26  Introduction  ..................................................................................................................................  26  Research questions  .......................................................................................................................  26  Purpose Statement  ........................................................................................................................  27  Positionality Statement  .................................................................................................................  27  Research Design  ...........................................................................................................................  28  Research Tradition  .......................................................................................................................  29  Participants  ..................................................................................................................................  30  Data Collection  .............................................................................................................................  35  Data Storage  .................................................................................................................................  41  Data Analysis  ................................................................................................................................  43  Trustworthiness  ............................................................................................................................  49  Limitations  ...................................................................................................................................  51  

Chapter IV: Report of Research Findings ................................................................................ 52  Research Questions  .......................................................................................................................  52  Teacher Assessment (Pre and Post Tests)  .....................................................................................  53  Student Assessment #1, Multiple-Choice Vocabulary Assessment  ...............................................  54  Student Assessment #2, Nonsense Word Completion  ...................................................................  60  

First Cycle Coding - Grade Four ................................................................................................ 68  First Cycle Coding - Grade Five ................................................................................................. 75  Second Cycle Coding - Grade Four ............................................................................................ 88  Second Cycle Coding - Grade Five ............................................................................................. 97  

Additional Data  ..........................................................................................................................  108  Grade Level Focus Groups  .........................................................................................................  108  

Growth/Achievement ................................................................................................................. 118  

       

   

3  

Difficulties/Challenges ............................................................................................................... 122  Fidelity/strategies ....................................................................................................................... 124  

Individual Teacher Interviews  ....................................................................................................  126  Growth/Achievement ................................................................................................................. 133  Difficulties/Challenges ............................................................................................................... 136  Fidelity/Strategies ...................................................................................................................... 138  

Chapter V: Discussion of Research Findings ......................................................................... 140  Introduction  ................................................................................................................................  140  Findings  ......................................................................................................................................  142  

Finding A: Fidelity and collaboration ...................................................................................... 144  Finding B: Strategic approaches .............................................................................................. 147  Finding C: Conceptual understandings ................................................................................... 149  Finding D: Student and teacher engagement .......................................................................... 151  

Implications for practice  .............................................................................................................  153  Recommendations  ......................................................................................................................  155  Further Research  ........................................................................................................................  159  Conclusion  ..................................................................................................................................  163  Action  .........................................................................................................................................  163  

References .................................................................................................................................. 165  Appendix A – Grade 4 cumulative assessment, lessons 1-5 ................................................................. 171  

Appendix B – Grade 5 cumulative assessment, lessons 1-5 ................................................................. 173  

Appendix C – Nonsense words for grades four and five assessments ................................................ 177  

Appendix D – Multiple-choice assessment for grade four ................................................................... 178  

Appendix E – Multiple-choice assessment for grade five .................................................................... 182  

Appendix F – Grade Four Student Assessment Questions by Type ................................................... 186  

Appendix G – Grade Five Student Assessment Questions by Type ................................................... 191  

Appendix H – Nonsense word assessment for grade four ................................................................... 196  

Appendix I – Nonsense word assessment for grade five ...................................................................... 200  

Appendix J – Focus Group Questions ................................................................................................... 204  

Appendix K – Individual Interview Questions ..................................................................................... 205  

Appendix L – Answer keys for multiple-choice assessments (Grades four and five) ....................... 206  

Appendix M – Answer keys for nonsense word assessments (First cycle coding for grades four and five) ........................................................................................................................................................... 207  

Appendix N – Nonsense word assessment cycle one coding for all grade four classes ..................... 208  

Appendix O – Nonsense word assessment cycle one coding for all grade five classes ...................... 213  

Appendix P – Updated nonsense word assessment after cycle TWO coding for all grade four classes ................................................................................................................................................................... 218  

Appendix Q – Updated nonsense word assessment after cycle TWO coding for all grade five classes ................................................................................................................................................................... 223

       

   

4  

Dedication  It is with utmost humility that I complete this journey since it would have never been possible

without my support network. I thank my advisers and mentors, Dr. Conn, Dr. Huizenga, Dr.

Green, and Dr. Keough for their guidance and support. I am also thankful to Mom C, my Maine

family and L3 for your love and encouragement. When it comes to my parents, Janice and

Dennis Brightman, words cannot express my love and admiration for you. Literally and

figuratively, I am who I am and I accomplish what I do because of you. Emma and Cody, you

are my reason for being. I appreciate all of your patience during this process. You are a daily

blessing in my life. The rest of my gratitude, and all of my heart, belong to Billy. You have been

my partner on many paths, and we have just completed another one…together. I wouldn’t have it

any other way.

       

   

5  

Abstract  Accessing vocabulary in the content areas requires different approaches than in the typical

English Language Arts classroom. In the sciences, there are at least three characteristics that

create the necessity to approach vocabulary differently; the novelty of the concepts the

vocabulary represents to students, the inordinate amount of new words encountered in the

sciences, and the lack of explicit vocabulary acquisition instruction provided by science teachers.

If students were able to leverage their understanding of words and word meanings and improve

their automaticity of understanding the morphology of words, they may be able to improve their

overall comprehension of text as well as concepts. The Matthew Effect and the Reciprocal

Model both underscore the causal nature of the vocabulary-reading comprehension relationship.

They also support the sense of immediacy necessary to keep children from falling further and

further behind their peers. It is possible, even at the intermediate level, to enhance a student’s

reading comprehension skills and content area conceptual understandings through direct

vocabulary study. This study documents the benefits of using Greek and Latin word parts in ELA

to acquire vocabulary and understand new concepts in science in grade four and five classes.

Students who participated in direct, explicit instruction of Greek and Latin word parts

outperformed students who did not in identifying unfamiliar science concepts that utilized those

word parts. Students indicated conceptual understanding of the scientific concept behind a

nonsense word due to their familiarity with a science-related morpheme, thus indicting an

opportunity to help students better understand new, content area vocabulary as well as the

concepts represented by that vocabulary.

KEY WORDS: vocabulary, science, conceptual understandings, morphology, literacy

       

   

6  

Chapter  I:  Introduction  

“Vocabulary is the glue that holds stories, ideas, and content together…making comprehension accessible for the child” (Rupley, Logan, and Nichols 1998/1999).

Statement of the Problem  

An increasing number of secondary students are entering high school unable to read at

grade level. In today’s easily accessible, information-rich global society, these students must be

able to understand and synthesize information from complex nonfiction texts much more quickly

and critically than ever before. With the release of the Common Core State Standards and the

subsequent Massachusetts Curriculum Framework Standards, it is clear that a significantly

greater focus has been placed on the importance of close reading of content area informational

text. Unfortunately, content area teachers, particularly at the secondary level, are traditionally ill-

prepared to assist students in utilizing literacy skills in their classrooms to acquire new

vocabulary and comprehend text and rarely teach these strategies explicitly in the context of their

subject matter.

Topic. The English language is comprised of several components. One of these is the

phoneme. A phoneme is the smallest unit of a language. An example would be the sound /p/ in

the word “pot.” A grapheme is a unit of a written system of language such as the letter “p.”

When phonemes are represented graphically, through graphemes, they develop meaning. The

smallest possible unit of meaning in a word or speech is a morpheme. Examples of morphemes

would be “pot” and “s” in the word “pots.” The morpheme “pot” identifies the cooking

implement while the morpheme “s” gives the meaning of plurality.

The most common form of intervention for struggling readers has traditionally been

phonemic in nature. Children work on mastering the sounds of letters, blends, etc. This is

       

   

7  

essential for successful reading. If a child does not have mastery of automaticity when decoding

text, they will expend far too much cognitive effort “sounding out” the letters and words making

it impossible to focus on the meaning of the words and text. However, once students have

mastered reading with fluency (speed, accuracy, and prosody), they may continue to experience

difficulty understanding the text they read. This is particularly common at the intermediate level

(grades four to five).

At this level, their cognitive effort may be compromised by an inordinate amount of

concentration on or misunderstanding of the morphemes. Therefore, a morphological approach,

or an explicit study of the smallest meaningful parts of words, may be beneficial in creating

“fluency” in their understanding of words. Accessing domain-specific vocabulary in the content

areas may require an even more concentrated approach since the vocabulary, again, most

prevalently at the intermediate level, becomes much more abstract and provides far less context

clues in the text to assist in decoding the meaning.

Problem statement. Accessing vocabulary in the content areas requires different

approaches than in the typical English Language Arts classroom. In addition, these approaches

will vary based on the unique complexities of each content area. In the sciences, there are at least

three characteristics that create the necessity to approach vocabulary differently. First, scientific

terminology is almost always used to identify concepts that are entirely new phenomena to

students and they are difficult to define with a synonym or comparative idea. Second, there is an

inordinate amount of vocabulary necessary in the study of science. It is most comparable to

learning a foreign language. Finally, the strategies that would perhaps be the most beneficial in

helping students to acquire new vocabulary are traditionally only taught explicitly by the English

       

   

8  

Language Arts teachers, and children do not tend to transfer these skills from one domain to

another on their own.

Science teachers report that they do not have the explicit understanding of the

fundamentals of literacy to assist students in acquiring the skills needed to compensate for a lack

of understanding when it comes to acquiring new vocabulary. There is little collaborative time

provided for teachers of English Language Arts and other content areas such as science to meet

and share their strategies for teaching vocabulary and other skills. In addition, they are largely

unaware of the strategies and terminology of their colleagues. Therefore, they are unable to make

explicit connections for the children based on what they may have learned as a strategy in

another class to transfer that knowledge in other content areas or from year to year. If teachers

used a coordinated system for introducing and reinforcing vocabulary across content areas and

across years, and if they worked collaboratively to reinforce not only this range of vocabulary

but the strategies used to understand the words and text they were reading, then students would

be able to leverage their understanding of words and word meanings and be able to improve their

automaticity of understanding the morphology of words and improving their overall

comprehension of text as well as concepts.

This indicates an “Action-Knowledge Conflict” as defined by Jacobs (2011). While

content area teachers understand the importance of students’ deep knowledge of domain-specific

vocabulary, they do not have the capacity to lead them through a morphological study of the

words and their word parts in order to increase their background knowledge and conceptual

understandings. This creates a “gap” and “prevent(s) individuals from behaving in professional

situations as they might wish otherwise” (Jacobs, 2011). This study documents the benefits of

using a coordinated study of morphology in vocabulary acquisition in grade four and five

       

   

9  

classes. It will describe how students who participated in direct, explicit instruction of Greek and

Latin word parts outperformed students who did not in identifying unfamiliar science concepts

that utilized those word parts. Students indicated conceptual understanding of the scientific

concept behind a nonsense word due to their familiarity with a science-related morpheme, thus

indicting an opportunity to help students better understand new, content area vocabulary as well

as the concepts represented by that vocabulary.

Justification for the Research Problem  

The first difficulty noted facing science teachers is the conceptual nature of the

vocabulary their students need to acquire. Unlike English Language Arts teachers, the

professionals who teach other content areas do not tend to incorporate a great deal of reading and

writing as part of their regular instruction. Mastery of conceptual understandings and reasoning

skills are best represented through these means, but content area teachers do not tend to use

language as a means of capacity building for their students (IES, 2008). For this reason, it is

difficult to accurately assess the extent of a student’s understanding. According to Halliday

(1993), measuring a student’s understanding of the “jargon” of science can be a challenge since

“making meaning is an internal function” and is best demonstrated through language (Cruz,

2012). If a student does not have the appropriate vocabulary to articulate their thoughts and

ideas, those thoughts and ideas are meaningless. Students need academic terminology and

experience with science vocabulary in order to effectively communicate about science (Lee and

Fradd, 1996; Lemke, 1990).

The second challenge to the acquisition of scientific vocabulary is the sheer mass of the

words necessary to ensure comprehension. Students will encounter over 180,000 words in

       

   

10  

course-related texts (Rasinski, 2008). According to James Wandersee (1985), “there are more

new words introduced in a beginning biology course than in the first-year study of a foreign

language.” Most science teachers do not have a systematic or multi-dimensional approach to

teaching such a vast amount of necessary words that will make them meaningful to their students

and to the concepts they will need to understand.

Finally, while content area teachers tend to realize the significance of knowing specific

vocabulary for their subject area, most fail to employ effective means and strategies to guarantee

their students’ understandings. It would be helpful for teachers, particularly science teachers, to

better understand language acquisition pedagogy as well as the application of scientific terms to

the context of the topics being taught (Carrier, 2005). While content area teachers are showing

more interest in incorporating literacy skills in their lessons, most do not have the skill set

necessary for doing this (D’Arcangelo, 2002). Incorporating targeted strategies for acquiring

scientific terminology will help to improve reading comprehension and conceptual

understanding. Teachers need to be provided with the appropriate supports necessary to help

them have a better understanding of word study and strategies for implementation (Bloodgood

and Pacifici, 2004).

Deficiencies in the Evidence  

The deficiencies in the research relate directly to the nature of acquiring vocabulary in the

content areas other than English Language Arts. There are many activities and strategies

available for the English teacher. However, many “experts” tend to recommend the use of these

same strategies in the other content areas and they are not always applicable. For example, dense,

non-fiction text such as a scientific article rarely provides context clues for the reader as in a

       

   

11  

work of literature. Therefore, simply providing these suggestions to science teachers will not

only be futile and ineffective, but they will be frustrating to both the teacher and the student who

attempts to utilize them. More specific information regarding the unique challenges in learning

science vocabulary, specifically the conceptual nature of the terms, the management of the

volume of words, and the strategies most applicable to scientific text, is needed to provide the

proper support for these teachers and their students.

Relating the Discussion to Audiences  

Content area teachers have expressed frustration in the way that traditional English

Language Arts teachers have suggested they approach content area literacy (Siebert and Draper,

2008). This study took into account the specific differences in approaching literacy, specifically

vocabulary acquisition, in the content areas. In particular, it focused on how this works in a

science classroom. This can be applied to other content areas in the sense that it will become

necessary to research the specific needs and challenges of each subject before implementing

strategies that may have traditionally worked in other fields. It also cannot be assumed that

teachers who are highly qualified in their content areas have the necessary pedagogy and

capacity necessary to help their students access the language needed for comprehension.

Providing this support can be beneficial to both students and educators.

For the intermediate school where this study took place, it is possible that the specific

methods used may suggest similar success in other subjects if the same care is given to research

the compatibility of the strategies, the content matter, and the nature of the subject. This could

potentially affect the way that students learn (and understand) not only the content, but the

concepts.

       

   

12  

Significance of the Research Problem: Implications of Poor Comprehension  

When considering science skills in particular, the Massachusetts Comprehensive

Assessment System (MCAS) scores show a lower level of proficiency and slower rate of gains,

particularly in middle school than other content areas. At the state level, students perform

considerably lower on the science MCAS than on English Language Arts or mathematics.

The implications of poor comprehension have an even greater effect on students who are

preparing for college and career readiness. For over 50 years, ACT has been reporting on student

readiness for post-graduate studies. In 2012, it was reported that only “67% of all ACT-tested

high school graduates met the English College Readiness Benchmark.” Just over 50% met the

Reading Benchmark, less than 50% met the Mathematics Benchmark, and “just under 1 in 3,” or

a mere “31% met the College Readiness Benchmark in Science” (ACT, 2012).

According to a study conducted by Achieve, Inc., students have not been appropriately or

sufficiently prepared for either college or career after high school. In college, 42% of instructors

said that their students are “not adequately prepared…for the expectations of college classes and

are struggling or having to take remedial courses to catch up” (2005). In the same study,

employers estimated that “39% of recent high school graduates…are unprepared for the

expectations they face at entry-level jobs…(and) 45% are not adequately prepared for the skills

and abilities they need to advance beyond entry level” (2005). While slightly under half of the

students polled felt that their high school education did not prepare them well for skills such as

research and effective communication, a full “51% felt there were gaps in their preparation in

science” (2005). These figures only considered the students who graduated from high school.

Many others never made it that far.

       

   

13  

In 2004, the Carnegie Corporation sponsored the Reading Next report. In it, authors, Gina

Biancarosa and Catherine E. Snow illustrate societal implications of lacking literacy

competency. For example, every day, close to 7,000 students drop out of high school (Biancarosa

and Snow, 2006). It is widely accepted that the majority of these students leave high school

because they cannot keep up with the literacy skills required to be successful. In fact, 70% of the

students entering grade 9 are reading below grade level (Biancarosa and Snow, 2004). This is

particularly concerning considering that the lowest 25% of the students in grade 9 are twenty

more times as likely to drop out of high school than their high performing peers (Carnvale,

2001). While “the 25 fastest growing professions have far greater than average literacy

demands…the 25 fastest declining professions have lower than average literacy demands”

(Barton, 2000). This, of course, leads to higher unemployment as “the demand for unskilled

labor decreases” (OECD, 2000).

Finally, as Massachusetts has adopted a new set of Massachusetts Curriculum

Frameworks for English Language Arts and Literacy, incorporating the Common Core State

Standards, there is a higher level of accountability for alignment. The standards include Literacy

in the Sciences and Technical Subjects. This was not always explicitly measured. According to

the National Science Education Standards, scientific literacy has many degrees and forms, but

included in the definition is being able to “display scientific literacy in many ways including

using technical terms or applying scientific concepts and processes” (National Academy of

Sciences, 1996).

In the English Language Arts and Literacy Massachusetts Curriculum Frameworks

adopted in March 2011, students in grades 9 and 10 are expected to “determine the meaning of

symbols, key terms, and other domain-specific words and phrases as they are used in a specific

       

   

14  

scientific or technological context” (MA DESE, 2011). As stated previously, direct instruction of

isolated and arbitrary terms is not effective. The new standards reflect this as well by adding that

students must also be able to “analyze the structure of the relationships among concepts in a text,

including the relationships among key terms” (MA DESE, 2011). This reflects the complexities

of vocabulary acquisition as well as its close relationship to understanding related text. Meeting

these standards will be essential at the local level in order to meet accountability standards.

Indeed, as the framework represents the standards of college and career readiness, all learners

must meet them for future success.

This focus on the importance of content area literacy in science is reflected in the most

recent versions of national and state science standards as well. In April 2013, an effort made by

several partnering states in collaboration with the National Research Council (NRC), the

National Science Teachers Association (NSTA), the American Association for the Advancement

of Science (AAAS), and others resulted in the release of the national Next Generation Science

Standards (NGSS) (Achieve, 2013). During a thorough review that will continue until an

anticipated completion in the 2015-2016 school year, the Massachusetts Department of

Elementary and Secondary Education released a draft revision of the NGSS in January 2014

(MA DESE, 2014).

Both of these documents focus on the ELA standards in relation to mastering the science

concepts and content. In fact, both the NGSS and the Massachusetts draft include “Connections”

boxes in every domain and grade level. These boxes indicate both the Mathematics and ELA

Common Core State Standards that are most closely related to and support the acquisition of the

science standards noted. This further reinforces the interdependent relationship between ELA

and science.

       

   

15  

The relationships and convergences among ELA and science as well as mathematics are

graphically represented in Figure 1.1. Tina Cheung of Stanford University provides an

explanation of this diagram. In it, Cheung states that the capacity that indicates students must

“demonstrate independence in reading complex texts, and writing and speaking about them” is a

summary of essential integrated ELA skills. Among these are skills that “attend to the language

demands expected from students” such as “comprehend and evaluate complex texts,…construct

effective arguments, and convey…information… discern a speaker’s key points, request

clarification, and ask relevant questions…build on others’ ideas, articulate their own ideas, and

confirm they have been understood… demonstrate command of standard English and acquire

and use a wide-ranging vocabulary” (Cheung, 2013).

Figure 1.1: Relationships and convergences in content area standards (2013).

       

   

16  

Research Questions  1. A. How do students apply their understanding of Greek and Latin word parts

in ELA to acquiring vocabulary and understanding new concepts in science?

B. What contributes to student application of morphological strategies learned in ELA to

their study of vocabulary in science?

2. A. To what extent do students utilize their understanding of morphemes to decode and

comprehend or conceptualize unfamiliar scientific vocabulary?

B. How do students hypothesize the meaning of nonsense scientific vocabulary

utilizing their understanding of Greek and Latin word parts?

Theoretical Framework  

There is a large body of research that supports the link between vocabulary and reading

comprehension (Chall and Jacobs, 2003; Beck, McKeown, and Omanson, 1987; Stahl and

Fairbanks, 1986; Graves, 1986; Nagy, 1988; Anderson and Freebody, 1981; NRP, 2000; Sinatra,

Berg, and Dunn, 1985; Baumann, Kame’enui, and Ash, 2003; Hiebert and Kamil, 2005;

Mezynski, 1983). While the evidence shows that the relationship is correlational, researchers

differ on whether the relationship is causational. However, Keith Stanovich (1986) highlighted

that researchers were beginning to examine these relationships differently. They began to

understand that the relationship between vocabulary development and reading ability might be

one of reciprocal causation. In the case of vocabulary acquisition, it was presented that

vocabulary ability is a determinant of reading ability (Stahl, 1983; McKeown, et al, 1983). Other

researchers also concluded that extensive reading has a significant influence on vocabulary

(Nagy, Herman, and Anderson, 1985).

       

   

17  

The Matthew Effect. Keith Stanovich presented the idea of reciprocal causation under

the examination of the Matthew Effect (1986). Originally coined in sociology by Robert K.

Merton in 1968, the Matthew Effect references the Biblical quote that is loosely translated to,

“the rich get richer and the poor get poorer,” and represents the cumulative advantage

phenomenon.

In the case of reading, as students successfully access the vocabulary needed to aid their

comprehension, this improves their understanding of text. They receive more positive feedback,

find pleasure in it, and continue to read more and more. This increases the amount of language

they are exposed to, and they acquire a richer vocabulary. As they gain a much broader exposure

to vocabulary, they become “richer.”

Conversely, the student who does not have easy access to text meets frustration. They do

not have a positive experience accessing vocabulary and do not understand the text. They fall

behind, become disengaged, and tend to read far less. This limits their opportunities to increase

their vocabulary and improve their reading skills. If they do not have the appropriate vocabulary

necessary to understand the text, they are likely to receive negative or even no feedback, which

in turn creates distaste for further reading. As they continue to avoid reading, they decrease their

access to varied text and their opportunity to learn vocabulary, becoming “poorer.”

This phenomenon is exacerbated as students enter intermediate school and are exposed to

increasingly more complex text with a decrease in the scaffolding and supports needed to access

that text, particularly in the content areas outside of English Language Arts, such as science. If

students do not receive the supports and structures they need to understand more difficult

vocabulary, the chances that their comprehension of expositional text will improve are

significantly minimized.

       

   

18  

Reciprocal Model. The seminal text in examining the relationship between vocabulary

and instruction was written by Anderson and Freebody in 1981. They suggested three models

through which to frame the correlation. The first is the Instrumentalist Model. This theory asserts

that having a larger knowledge of vocabulary increases understanding of text, or “knowing the

words enables text comprehension.” If this is accurate, it can be hypothesized that teaching more

vocabulary words would have a direct effect, or causation, on successful comprehension. The

second theory is the Knowledge Model. This acknowledges the role of yet a third variable:

background knowledge. Therefore, having an understanding of the subject matter facilitates

comprehension of the text. The vocabulary that the individual has mastered is “merely the

exposed tip of the conceptual iceberg.” The third model is called Aptitude. This posits that

subjects with a large vocabulary “possess superior mental agility,” giving them a natural capacity

for text comprehension.

Many researchers, including Anderson and Freeman, assert that the true model of the

vocabulary-comprehension may not be only one of these models at the exclusion of the other

two, but instead some combination of them. “No serious scholar in reading or related fields

adheres to any one of these positions” (Anderson and Freebody, 1981). Subsequent theories in

various combinations thus were presented from researchers such as Karen Mezynski (1983) and

Stephen Stahl (1991).

However, in 2005, William Nagy most recently suggested, similar to the Matthew Effect,

that the causal relationship between vocabulary and comprehension was more bidirectional in

nature. He suggested the Reciprocal Model. “Having a big vocabulary does contribute to being a

better reader. But being a good reader also contributes to having a better vocabulary.” He further

recommends creating a comprehensive literacy program with distinct characteristics. These

       

   

19  

include “teaching individual words; extensive exposure to rich language (both oral and written);

and building generative word knowledge” (Nagy, 2005).

The Matthew Effect and the Reciprocal Model both underscore the causal nature of the

vocabulary-reading comprehension relationship. They also support the sense of immediacy

necessary to keep children from falling further and further behind their peers. It is possible, even

at the intermediate level, to enhance a student’s reading comprehension skills through direct

vocabulary study. However, this must not result merely in a drilled practice of isolated word

banks. Determining which strategies are the most effective will be essential. This necessitates

careful consideration of the content area to be analyzed as well as the nature and context of the

instruction.

       

   

20  

Chapter II: Literature Review

It is clear that many studies and researchers have concluded that vocabulary is directly

connected to comprehension. It is further evident that explicit strategies regarding language

acquisition, particularly in content areas like science, would be the key in getting students to

understand and better articulate scientific concepts. However, the research is inconclusive about

what specific strategy or strategies are the most effective in creating this increase in vocabulary

and comprehension (NICHHD, 2000; Biancarosa and Snow, 2004). Most researchers would also

agree that vocabulary instruction alone is not sufficient (Graves, 1986; Beck, Mckeown, and

Kucan, 2002; Mezynski, 1983; Stahl, 1991), which makes the strategies used that much more

important to carefully consider and further research.

Scarborough’s Reading Rope Model of Reading  

A commonly used model of reading today is Scarborough’s Reading Rope (2001) (Figure

2.1). This visual representation of the multi-faceted nature of reading separates word recognition

and language comprehension. Phonological awareness, decoding (and encoding, or spelling), and

sight recognition are parts of “word recognition.” “Language comprehension” is made up of

background knowledge, vocabulary knowledge, language structures, verbal reasoning, and

literacy knowledge. The individual strands combine under the two broader strands, which then

become intertwined to represent the skill of reading. Under this model, vocabulary is represented

as a necessary component for comprehension.

       

   

21  

Figure 2.1: Scarborough’s Reading Rope (2001).

In 2000, the National Reading Panel published its report “Teaching Children to Read.” In

it, the NRP identified five components that are essential for students learning to read effectively

(NICHHD, 2000). Three of these components, phonemic awareness, phonics, and fluency (which

would be represented under “word recognition” in Scarborough’s Rope) have long been

considered “mastered” in the primary grades. Vocabulary and reading comprehension, the

remaining two components, have been seen as ongoing throughout a child’s academic career.

Unfortunately, this is not always the case. In fact, there is an increasing population of

students that cannot effectively access complex text in the intermediate grades and subsequently

into high school. Further, it is becoming increasingly clear that the way to approach vocabulary

and comprehension are different depending on the specific content area.

       

   

22  

Reading comprehension, the pinnacle of the five components is inextricably linked to

vocabulary acquisition. Many researchers and studies have supported the connection between

vocabulary and reading comprehension (Chall and Jacobs, 2003; Beck, McKeown, and

Omanson, 1987; Stahl and Fairbanks, 1986; Graves, 1986; Nagy, 1988; Anderson and Freebody,

1981; NICHHD, 2000; Sinatra, Berg, and Dunn, 1985; Baumann, Kame’enui, and Ash, 2003;

Hiebert and Kamil, 2005; Mezynski, 1983). Domain-specific vocabulary is essential when

studying any subject matter, and “reading vocabulary is crucial to the comprehension process of

a skilled reader” (NICHHD, 2000). For this reason, it is important to explore the best means of

acquiring both fundamental and complex vocabulary in the content areas.

While vocabulary acquisition is critical to effective comprehension in all areas, this is

particularly true in the content areas such as science. This text is largely expositional, there are

few context clues to develop meaning, the terms are conceptually vs. semantically-based, the

language of the discipline is growing at an exponential rate, and content area teachers are

traditionally not experienced at providing strategies for decoding complex terms. For students

who already have difficulty accessing this type of text, this deficit puts them at a distinct

disadvantage for performing independent tasks and succeeding in the sciences.

Conceptual nature of domain specific science vocabulary  

Content area vocabulary is largely conceptual in nature vs. semantic. A reading teacher

can make clear connections via a student’s common understanding of other vocabulary words

(i.e.: “melancholy” means very sad), and the student can make the link to prior knowledge or

experience of sadness. However, for a science teacher attempting to make a similar connection

(i.e.: “photosynthesis”), it would be very difficult. There is an underlying concept which the

       

   

23  

student has not yet mastered. This makes the learning of this vocabulary a far more complex

endeavor.

The study of Greek and Latin word parts is particularly significant in the acquisition of

science terms and concepts. Much of the new vocabulary that students will encounter, especially

in science, will be based on Greek and Latin (Partnership for Reading, 2001). Students can gain

far better understandings of science concepts if they already have a conceptual understanding of

the roots of the terms (Aaron, Joshi, and Quatroche, 2008).

Extent of necessary vocabulary  

In order for a student to comprehend the material they are reading, they must have an

understanding of 90 to 95 percent of the words in the text (Hirsch, 2003). With the increasing

rigor of the texts students are expected to read, they will need to learn 2,000 to 3,000 new words

per year in order to understand them (Beck, McKeown & Kucan, 2002). However, the average

student can learn approximately 400 words per year through direct instruction (Beck, McKeown

& Kucan, 2002). Teachers in subjects with a constantly growing and evolving pool of knowledge

and language, like science, must use a different approach.

There are approximately 1.2 to 2 million words in the English language, and technology

is adding at least 20,000 words to that list each year. Even though a student will be expected to

learn up to 3,000 words each year just to maintain grade level vocabulary, they can only master

8-10 per week through direct instruction (Beck, McKeown, and Kucan, 2002). This will not be

sufficient. However, if students study the morphology of words instead of isolated vocabulary,

they can increase the number of words they learn exponentially. Even learning the more common

affixes and bases can assist students in understanding many new words (Partnership for Reading,

       

   

24  

2001). Of English words with more than one syllable, 90% are Latin based, and most of the

remaining 10% are Greek based (Rasinski, 2008). From one Latin root, a student can decode or

generate five to twenty other words (Rasinski, 2008). This will help them to access the words

(more than 180,000) that they will encounter in school texts.

Need for strategies  

In June 2012, Massachusetts adopted a new set of English Language Arts Curriculum

Frameworks based on the Common Core State Standards. Included in it are standards for

Literacy in the Technical Subjects. The standards illustrate the need for students to be able to

successfully read scientific and technical texts. They must be able to “read complex

informational texts in these fields with independence and confidence because the vast majority of

reading in college and workforce training programs will be sophisticated nonfiction” (MA

DESE, 2012). Beyond simply memorizing new terms, “the vocabulary standards focus on

understanding words and phrases, their relationships, and their nuances, and on acquiring new

vocabulary, primarily general academic and domain-specific words and phrases” (MA DESE,

2012).

In the primary grades, teachers focus on foundational reading skills because students are

“learning to read.” Beginning in the fourth and fifth grades, the emphasis in literacy is now

“reading to learn” by providing the strategies that help them learn to access text and appreciate

author craft. Traditionally, both the “learning to read” and “reading to learn” strategies have

focused on the use of narrative and fictional text. Content area text, however, is expositional in

nature. Students have had far less experience decoding the vocabulary and making meaning of

dense expositional text by the time they enter high school. Just at the time when explicit reading

       

   

25  

instruction ends, students are being asked to engage with far more rigorous text. Strategies that

students had been encouraged to use in the past, such as utilizing context clues, will not be as

helpful. Most of this text makes little use of cues in context.

Science teachers have not traditionally embraced teaching literacy strategies. However, in

order for students to be critical readers of complex expositional scientific text, they will need

explicit strategies for accessing the language in order to better comprehend the content, and they

will not be able to rely on the most common strategies used when they learned to read. Since

researchers have also determined that there is not one strategy that best meets this need

(NICHHD, 2001; Biancarosa and Snow, 2004), it is important to investigate elements that may

prove successful.

       

   

26  

Chapter III: Research Design

Introduction  

Accessing vocabulary in the content areas requires different approaches than in the

typical English Language Arts classroom. In the sciences, there are at least three characteristics

that create the necessity to approach vocabulary differently: the novelty of the concepts the

vocabulary represents to students, the inordinate amount of new words encountered in the

sciences, and the lack of explicit vocabulary acquisition instruction provided by science teachers.

If students were able to leverage their understanding of words and word meanings and improve

their automaticity of understanding the morphology of words, they may be able to improve their

overall comprehension of text as well as concepts. The Matthew Effect and the Reciprocal

Model both underscore the causal nature of the vocabulary-reading comprehension relationship.

They also support the sense of immediacy necessary to keep children from falling further and

further behind their peers. It is possible, even at the intermediate level, to enhance a student’s

reading comprehension skills through direct vocabulary study.

Research questions  1. A. How do students apply their understanding of Greek and Latin word parts

in ELA to acquiring vocabulary and understanding new concepts in science?

B. What contributes to student application of morphological strategies learned in ELA to

their study of vocabulary in science?

2. A. To what extent do students utilize their understanding of morphemes to decode and

comprehend or conceptualize unfamiliar scientific vocabulary?

B. How do students hypothesize the meaning of nonsense scientific vocabulary

utilizing their understanding of Greek and Latin word parts?

       

   

27  

Purpose Statement  

The purpose of this basic qualitative study was to discover whether the use of

morphology skills was transferred and utilized in the understanding of science terms and

concepts by fourth and fifth grade students at a suburban Massachusetts intermediate school to

improve their comprehension. Morphology skills were generally defined as combining and/or

substituting known Greek and Latin affixes explicitly taught in the English Language Arts

classes in unfamiliar or even nonsense patterns to create meaning or a reasonable hypothesis of

meaning.

Positionality Statement  

As a former high school English teacher, I have an interest in language and linguistics. I

have not taught a science course, but I have an appreciation of the fact that it is a conceptual and

inquiry-based subject, and one in which ideas, procedures, and processes must be articulated

clearly and succinctly. In order to do this, it is important to have a firm grasp of the language

required to be accurate and direct.

I have observed many science colleagues and protégés, and I have found that there is a

lack of direct vocabulary instruction that is meaningful and effective, and there is a lack of

appreciation for indirect vocabulary study altogether. While I may be biased in the belief that

better instruction will enhance student understanding of the overall concepts, I did not have a

bias as to which strategy or strategies would be most successful, particularly in a science setting.

In practice, a morphological approach has been helpful to students in accessing a broad range of

vocabulary words in my English courses. However, I am well aware that strategies do not always

transfer well across content areas. While I have an interest in evaluating the effectiveness of the

       

   

28  

specific strategies utilized in a science class, I did not have a bias in terms of whether these

strategies will be as successful as they are in an English classroom or whether a particular

strategy is more effective than any other. Therefore, this limited my potential bias as to whether

the study of Greek and Latin word parts was or was not a successful strategy in assisting

comprehension of science concepts in this content area.

I acknowledge, however, that there may be some level of bias in other areas, but suggest

that this was beneficial to the research. Due to the focus of my education, my extensive

experience as a high school English teacher, and my background as a professional development

provider of PK-12 teachers, I brought a lens to the study that was useful. This study was

designed to determine if students were applying their morphological skills to their understanding

of science. Since this was determined to be the case, it was further necessary to determine to

what extent this was happening and to what level of comprehension the students had mastered a

vocabulary word. My understanding of both the content area and pedagogy was helpful in

determining that the students showed conceptual understanding and that the strategies used by

the classroom teachers were influential in this understanding taking place.

Further, having been a classroom teacher for twenty years, and having previously taught

in the district in which the study took place (though no longer employed there), there was a sense

from teachers that I understood their challenges and choices. It helped to “enhance the rapport

and dialogue” with the teachers (Ponterotto, 2005).

Research Design  

This research was designed as a basic qualitative study (Merriam, 1998). The research

interest was in “understanding the experience” (Merriam, 2009) of the participants, or, in this

       

   

29  

case, whether students made use of the language skills learned and utilized in their English

Language Arts classes to make sense of the language they encountered in science class.

According to Merriam (2009), “Qualitative researchers are interested in understanding the

meaning people have constructed, that is, how people make sense of their world and the

experience they have in the world.” In order to know if the strategies used in the English

Language Arts classes had implications beyond that subject area, it was necessary to determine

to what extent and to what level of success students had transferred this skill.

The paradigm of the study is “constructivism/interpretivism” as described by Ponterotto

(2005). From a constructivist perspective, reality is “subjective and influenced by the context of

the situation, namely the individual’s experience and perceptions, the social environment, and

the interaction between the individual and the researcher” (Ponterotto, 2005). Rich, thick

description was necessary in order to provide legitimacy to the study. The analysis is so

contextually specific, that another researcher may interpret it otherwise. This is especially true if

there is a desire to replicate the study in other school districts, different grade levels, or systems

with varying demographic populations.

Epistemologically, it is the intention of the study that “through intense interaction and

dialogue, both the participant and the researcher will reach deeper insights (hermeneutical

discovery)” of the experiences of the teachers and the students (Ponterotto, 2005). This can then

be useful in determining whether the use of the Greek and Latin word parts should or should not

be continued with or without modifications in the ongoing vocabulary instruction of the English

Language Arts classes.

Research Tradition  

       

   

30  

The study was conducted as a General Inductive Analysis. There was not one specific

theory or hypothesis being investigated in this study; rather, the intended outcome was “the

development of categories into a model or framework that summarizes the raw data and conveys

key themes and processes,” making decisions about the most important information, and even

taking into account the possibility of unplanned outcomes (Thomas, 2006).

This study aimed to determine whether or not students transferred their skills from one

area to another and what result this had on their comprehension. Using general inductive analysis

“provide(d) a simple, straightforward approach for deriving findings linked to focused evaluation

questions” (Thomas, 2006).

Participants  

The population of the sample included approximately 400 grade four and five students at

a suburban Massachusetts intermediate school and their teachers. The students ranged in age

from eight to eleven and included both boys and girls. This study employed purposeful selection

of participants (Maxwell, 2005). Also referred to as “purposeful sampling” (Patton, 1990) and

“criterion-based selection” (LeCompte and Preissle, 1993), Maxwell defines this as “a strategy in

which particular, settings, persons, or activities are selected deliberately in order to provide

information that can’t be gotten well from other choices” (2005).

In this case, the unique transition from “learning to read” to “reading to learn” that

converges with the reduction in targeted reading skills provided in class and the separation of

teachers for different content areas made the intermediate grades particularly appropriate for this

study. Maxwell quotes Weiss (1994) as adding that the sample (in this case, the teachers), is

more accurately coined a “panel,” because they are “people who are uniquely able to be

       

   

31  

informative because they are expert in an area or were privileged to witness an event” (Maxwell,

2005). The participating teachers, and indeed the students, reported on what they witnessed

themselves. This information was augmented with archival assessment information provided by

one of the teachers and the results of two student assessments created by the researcher.

Recruitment and access. Permission to conduct the study was secured, in writing, from

the Superintendent of the school district after a public presentation was made to the full School

Committee. Participants were fourth and fifth grade students and their teachers in the

intermediate school. Parents and teachers received a full briefing, in writing and orally, of the

scope and purpose of the study as well as a specific description of all of the data that would be

collected. In addition, two informational sessions were also be offered outside of regular school

hours (directly after school and again in the evening) and on school grounds for any parents

and/or teachers who wanted to meet in person and ask questions or secure clarification before

consenting to participate.

Students were only be asked to participate if their teacher explicitly taught vocabulary

throughout the year, provided pre and post assessments for these units, and agreed to administer

the study assessments of unfamiliar science vocabulary and nonsense science vocabulary (these

assessments had no impact whatsoever on student grades for this class).

The only reason for student exclusion, other than not being enrolled in one of these

classes, was if the student and/or parent/guardian declined participation. In this case there were

no consequence, and the child simply continued with their regularly scheduled school activities.

Students were not be identified by ethnicity/race, socio-economic level, literacy level or health.

The study did not include any non-English speaking students, parents, or teachers, as there were

       

   

32  

currently none identified at this school. All teachers who met the above criteria who elected to

participate were included.

The particular intermediate school chosen for participation, referred to as the Billings

Intermediate School, was selected because of their recent adoption of a morphological approach

to vocabulary instruction. Each teacher had been teaching the same grade level for several years.

In addition to their current experience using the new approach to vocabulary, they had historical

data that they could reference to reflect on any changes in student behavior and/or achievement

when compared to vocabulary study without this direct instructional approach.

Billings Intermediate is part of the Brighton-Cabot Regional School District. The sending

communities are both suburban, middle class towns that are comparable in land area and

demographics. The population of the two towns was approximately 9,076 and 10,623 and over

95% white as of 2009. The median household income for these communities as of 2011 was

$79,293 and $86,514 (Advameg, Inc.). The regional school district population as of 2013 was

3,055 with 493 of these students attending the Billings Intermediate School in grades four and

five. The number of students in the school who are designated as having Limited English

Proficiency, low income, or special needs is comparable to the other schools in the district, but

both the school and district are below the Massachusetts average for these indicators (MA DESE,

2014). There were no students at this school identified as ELL during this study.

As stated previously, Billings Intermediate serves all students in the district in grades four

and five. There are ten teachers for each grade level, not including special education teachers or

specialists. The ten classroom teachers work in paired teams and teach the same students. One

teacher is responsible for math and science and their partner colleague teaches English Language

       

   

33  

Arts and social studies. For purposes of this study, all classroom and special education teachers

were asked to participate since half of the classroom teachers teach ELA and half teach science.

The special education teachers work with students in both of these areas, they are present in most

classrooms while lessons are being conducted, and they teach small groups of students as well.

Each of these professionals observed students working with the vocabulary words addressed in

the study and had insights into their level of usage, understanding, and/or skill transference.

At the conclusion of the study, teachers who took part received a $10 gift card to Barnes

and Noble Booksellers. Students with permission who took part in the study received a small

school supply item (pencil, pen, eraser, etc.) As the school has been identified as a “latex-free

zone,” none of the supplies provided contained latex or latex products.

Protection of human subjects. The researcher presented and explained, in detail, the

roles of the students and teachers as well as the assessment information that would be collected

to the teaching staff who then explained this information, in grade appropriate language, to their

students. This way, the teachers all received the same information at the same time, and the

students had this information relayed to them by a familiar adult in their regular classroom

setting. Students had the opportunity to present any questions either orally or in writing and have

them answered to their satisfaction before participating. No students exercised this option.

Only teachers who agreed, in writing, to participate, and only students who assented and

whose parents did not elect to opt out, in writing, for their children to participate were be asked

to take part in the study. All participants were reminded of their right to withdraw at any time

during the course of the study and no data was used to identify individual students or teachers.

All names, including that of the school and district, remained anonymous as are not necessary to

publish for the scope of this study.

       

   

34  

A gatekeeper at the intermediate school provided information regarding the names of

students, their parents, and their addresses to the researcher. This information was used only for

disseminating the above information about the study to parents and informing them of the

informational sessions that were provided and their right to opt out of participation.

Confidentiality was maintained and all contact information was returned to the gatekeeper at the

conclusion of the study.

While careful note-taking was done during all interviews, any that were conducted in

person were recorded and transcribed. All participants were asked for permission before a

recording device was used. The recording was done for the purpose of transcription. A

professional transcriptionist was hired for each recording. The interviewer and the subjects, to

ensure accuracy in language, context, and intention, were then provided with full, unedited

copies of the transcripts and asked to comment or correct as necessary. None of the participants

made a correction, nor did anyone contest any of the content or context. A reputable,

professional transcription service was employed to maintain subject confidentiality (Rev.com).

All recordings will be destroyed at the conclusion of the successful defense of the study.

Losses of confidentiality of student scores or teacher interviews were the only risks

associated with this study. However, student scores of the researcher-administered assessments

did not impact their class score or average in any way. Also, the researcher identified classes and

students by number only in the coding process, and only the researcher has the key to which

numbers are assigned to each class. The key will be destroyed once the defense is complete, and

no student names or unique identifying details were included in the final product. Therefore, the

risk to students was minimal.

       

   

35  

The teacher interviews were conducted both in small groups (focus groups) and

individually. Teachers could elect to participate in either or both of these forums. The study

focused on student use of internalized strategy transfer as well as individual strategies used by

teachers and not on the ability or teaching competency of the teachers. Also, no teacher names or

unique identifying details were included in the study. Therefore, the risk to teachers was also

minimal.

Data Collection  

Data was collected through published assessments, two researcher-developed student

assessments, teacher focus groups, and teacher leader interviews. The results of this research are

a combination of the data collected and an analysis based on the research questions that focus on

meaning and understanding.

Collection: Teacher assessments. The instruments used in this study included published

assessments as well as researcher-developed tools for students to complete. The published

assessments include the activities that teachers administered this year from the text Growing

Your Vocabulary: Learning from Latin and Greek Roots A and B (Moliken,ed, 2008) published

by Prestwick House. The ELA Curriculum Leader shared the pre- and post- test scores for

students in two representative classrooms in grade five. This was in the form of the cumulative

reviews that are presented at the conclusion of every five units of study for grade four (Appendix

A) and grade five (Appendix B). Two of these five-unit assessments were provided for review

and comparison. This text was purchased and added to the ELA curriculum at Billings

Intermediate during the 2012-2013 school year. Students in grades four through twelve currently

use this vertical program for vocabulary instruction. The scores are counted as part of the

       

   

36  

students’ ELA grades for each quarter. These assessments are provided in Appendix B. The

results of this archival data were reviewed without student names for purposes of anonymity. For

this review, it was not necessary to know the identity of each child, so this information was

eliminated, and only the data was used for comparative purposes.

Collection: Researcher assessments. In order to determine if the students were

transferring their understanding of the vocabulary word parts they learned in ELA to their

science classes, the researcher developed two instruments to measure their level of conceptual

understanding of science terms. These term prompts included (1) words with word parts that the

students have learned as part of an explicit vocabulary unit in ELA and also as part of their

science curriculum, (2) words with word parts that the students have learned as part of an explicit

vocabulary unit in ELA but not part of their science curriculum, and (3) words with word parts

that the students have not learned as part of an explicit vocabulary unit in ELA but were part of

their science curriculum.

The purpose for the different categories of words and word parts was to determine a

possible causal factor in the event that students showed a greater or lesser level of understanding

depending on whether the Greek and Latin word parts were part of the student’s explicit

vocabulary program.

This study utilized prompts that focused on the highest levels of intellectual processes as

noted in the updated version of the taxonomy as well as the “hierarchical arrangement of

cognitive processes” postulated by Herbert Klausmeier (Frayer, Fredrick, and Klausmeier, 1969).

The first assessment was a multiple-choice questionnaire (found in Appendix D for grade

four and Appendix E for grade five). Each question on this measure was generated from

prototypes developed specifically for measuring student conceptual understanding (Frayer,

       

   

37  

Ghatala, & Klausmeier, 1972, Frayer, Fredrick, Klausmeier, 1969). The questions provided

further scaffolding to more specifically gauge the level of the students’ concept mastery based on

the research done at the University of Wisconsin resulting in “A Schema for Testing the Level of

Concept Mastery” (Frayer, Fredrick, Klausmeier, 1969).

Questions on the first probe of science vocabulary were constructed with the

understandings defined by Frayer, Fredrick, and Klausmeier (1969) to include concept instances

which were comprised of attributes which were relevant or irrelevant to the concept, non-

examples lacking one or more of these characteristics, concepts that had names and could be

defined “structurally, semantically, operationally, or axiomatically,” and concept instances that

were related to supraordinate, coordinate, and subordinate concepts. A breakdown of the type of

each question is found charted in Appendix F for grade four and G for grade five.

It has widely been recognized that the hierarchy of intellectual behaviors developed by

Benjamin Bloom and a team of educators in 1956 represents a scale that classifies the cognitive

domains important in acquiring knowledge. They are organized from the lowest order processes

to the highest: Knowledge, Comprehension, Application, Analysis, Synthesis, and Evaluation

(Bloom & Krathwohl, 1956). While it was originally hypothesized that the “lower” levels were

necessary to be mastered before moving on to the “higher” levels, this study focused on the more

recently accepted understanding that more than one cognitive process may be employed on a

particular task. In addition, the higher levels of cognitive processes were the focus of the student

prompts in order to identify true understanding of the concepts behind the words.

In 2001, the taxonomy was updated to reflect the skills necessary for success in the 21st

century. The terms for each of the processes were adapted to the verb form vs. the noun form,

and the highest two levels of the cognitive domain were inverted. In order of hierarchy, these

       

   

38  

processes are now: Remembering, Understanding, Applying, Analyzing, Evaluating, and

Creating (Anderson and Kathwohl, 2001). Whenever possible, the command verbs representing

the higher demands of cognitive domain were utilized in the assessments.

“A Schema for Testing the Level of Concept Mastery” (Frayer, Fredrick, Klausmeier,

1969) also identified two types of assessment items; Selection Items and Production Items. The

Selection Items require students to choose an answer based on a bank of selections provided.

This is the type of question that was used in the first researcher assessment. Five options were

provided for students for each answer bank. The Production Items prompt students to produce or

create their own responses and are more open-ended. The researcher found that the Production

Items better incorporated the language of Bloom’s Taxonomy of Educational Objectives and

offered the opportunity to show further levels of student competency and mastery. This was the

type of prompt that was used in the second assessment, the nonsense word questionnaire to

decrease the potential that students would choose an answer randomly based on a pool of

suggested responses. The assessments also contained the command verbs and prompts that were

most representative of the Bloom’s Taxonomy of Educational Objectives (Bloom & Krathwohl,

1956).

The nonsense word assessment provided a series of compound words that students had no

prior experience reading or hearing. One part of the word was a morpheme that was part of their

ELA curriculum, their science curriculum, or both their ELA and science curricula. These

morphemes were chosen carefully from the curriculum guides provided by the ELA Curriculum

Leader and the science Curriculum Leader. A word part was chosen if it was a Greek or Latin

root, if it was part of a word that was in the written curriculum for ELA, science, or both, and if

it was a science-related concept. Only morphemes that met all three criteria were considered.

       

   

39  

(Again, the complete bank of word parts chosen is in Appendix C). The other half of the

nonsense compound word was a simple word with which they were familiar (ie: book, pants).

Students were informed that they are working with nonsense words. They were then prompted to

provide a written description of the word and a drawing to represent the word pictorially. The

prompts were organized so that students could show their ability to apply their understanding of

morphemes and their uses in constructing new words to make meaning of familiar, unfamiliar, or

nonsense science terms. The nonsense word assessments are found in Appendix H for grade four

and Appendix I for grade five.

These tools were vetted and reviewed by educational colleagues outside of the Brighton-

Cabot Regional School District. Teachers of both fourth and fifth grade students were asked to

review the content of the assessment and comment on whether they considered it

developmentally appropriate for their students. They were further requested to note whether they

found the tool to be an accurate assessment of student understanding of the concepts. In both

cases, each teacher (four in total) agreed that the assessment was appropriate for the research

purposes and age/grade level of the target population.

The researcher-created assessments were administered by the science teachers in the

science classes during their regularly scheduled period and took no longer than 60 minutes to

complete. All science teachers were given instructions for the administration of the assessments

so that they were distributed and collected using the same procedure, including the directions to

the students. Having the students complete the assessments in the science classroom with their

science teachers eliminated their perception that this was an ELA task. It was also the intention

that students felt at ease due to the fact that they were participating in their regularly scheduled

class with a familiar adult. The assessments were also examined to see if students utilized any

       

   

40  

written note-taking or test-taking strategies or provided indications that they were reflecting,

metacognitively, on the skills they used to determine the meanings of the words or word parts

they were presented with in the instruments.

Collection: Teacher focus groups and interviews. The researcher developed teacher

focus group questions (Appendix J) and teacher leader (Curriculum Leader) interview questions

(Appendix K), which were used during each session. Focus group and interview protocols were

closely followed, and confidentiality was respected in all cases.

There were four Focus Groups; one for the grade four ELA teachers and special

educators working in these classes, one for the grade four science teachers and special educators

working in these classes, and two grade five groups of the same subjects. The reason for

separating the two content areas was so that the pedagogy used in each domain was not

suggested upon the other. For example, a coordinated scope and sequence of Greek and Latin

word parts was adopted in the ELA department. It is important to know whether or not this was

being used, to what extent, and with fidelity to the program or not in the science classes. If the

science teachers were not present when the ELA teachers discuss these procedures and strategies,

yet they are able to discuss them on their own, then the researcher could be more confident that

they were, in fact, utilizing these practices. This is the same reason the grade four and grade five

teachers participated in separate groups. These units of study were part of the required

curriculum. Interviewing the teachers separately provided a more accurate assessment of the

fidelity of the program in each grade, and teachers in one grade level or content area were not

able to influence the answers of the other grade level or content area.

The focus groups were conducted in main areas of the building. These common areas

provided a comfortable amount of space, were familiar to the participants, and were convenient

       

   

41  

for them to attend. The meetings took place during their regularly scheduled common planning

time. This was at the preference of the teachers and their building administrator. The groups

were designed to last for approximately 45 minutes; however, the interviewer left the group with

the opportunity for contact should they think of ancillary information or examples that they

wanted to share after the meeting. None of the participants exercised this option.

Individual teacher interviews were conducted with teacher leaders, called Curriculum

Leaders, in this district. Each of the content area team representatives (math/science and

ELA/social studies) has a peer leader for curriculum and assessment purposes. These Curriculum

Leaders run grade level meetings and attend district curriculum meetings regularly. They not

only had the experience of using these tools in the classroom, but they had the ability to

comment on which teachers were also using them and the level of success or challenge that they

were experiencing as well as the results of the classroom assessments for their students. In

addition, they participated in “learning walks” throughout the building, and directly observed the

practices in their colleagues’ classrooms.

All teachers also had the option of participating in an individual interview. Individual

interviews were approximately 40 minutes. The interviews took place in the teacher’s classroom

for the convenience of the teacher. This was during their preparatory period during the school

day. The intent was to make the setting and time most convenient for the subject. Interviews

were also available for individual teachers in the event they did not feel comfortable speaking in

a group setting and preferred to meet individually for any reason. Again, no teachers took

advantage of this option.

Data Storage  

       

   

42  

The researcher stored the data in hard copy and, where appropriate, electronically on a

home computer and flash drive. All information, including the key to individual assessment

scores and all transcripts, will be destroyed upon successful defense. All letters submitted to opt

out of participation will be kept in original hardcopy for three years, at which time the

documents will be destroyed.

Only the researcher and confidential transcriptionist have access to the audio recordings.

All participants were asked for permission before a recording device was used. The recording

was done for the purposes of transcription only and were in the form of an MP3 file recorded on

an iPhone 4 utilizing the Rev Voice Recorder: Audio Transcription app. This file was sent to Rev

Transcription Service. This company was selected due to their 98% accuracy rate and

commitment to confidentiality. The files were transmitted utilizing 128-bit SSL encryption for

security. The final transcription was in Microsoft Word format. All recordings will be destroyed

upon successful defense to maintain subject confidentiality.

Loss of confidentiality of student scores or teacher interviews was the only risk

associated with this study. However, student scores of the researcher-administered assessments

did not impact their class score or average in any way. Also, the researcher identified students by

number only and only the researcher has the key to which number was assigned to each student.

The key will be destroyed once the study is complete and no student names or unique identifying

details were included in the study. Therefore, the risk to students was minimal.

The teacher interviews were conducted both in small groups (focus groups) and

individually. Teachers were able to elect to participate in either or both of these forums. The

study focused on student use of internalized strategy transfer and not on the ability or teaching

       

   

43  

capacity of the teachers. Also, no teacher names or unique identifying details were included in

the study. Therefore, the risk to teachers was also minimal.

Data Analysis  

As Creswell states, the researcher in a basic qualitative analysis, “collects qualitative

data, analyzes it for themes and perspectives, and reports 4-5 themes” (2009). This was the case

in this inductive study.

Analysis: Teacher assessments. Students were already given unit activities and

assessments throughout the school year in their ELA classes as part of the vocabulary

curriculum. The Curriculum Leader for ELA, a grade five teacher, compiled longitudinal data of

the Cumulative Review assessments provided in the vocabulary text, Growing Your Vocabulary:

Learning from Latin and Greek Roots for two grade five classes. This data was analyzed to see

how the students performed in each set of five units in comparison to one another. Specific focus

was directed to whether the students showed significant improvement from pre to post

assessments, but also if they showed improvement from one pre-test to the next. This suggested

an increase in their overall word attack skills for unknown words.

Analysis: Researcher assessments. An important component of the researcher

assessment tool was its ability to determine not only whether a student could define a particular

word, rather, whether or not they had mastery of the concept behind the word. For this reason,

the assessment was developed using “A Schema for Testing the Level of Concept Mastery”

(Frayer, Fredrick, Klausmeier, 1969).

Questions on the first probe of science vocabulary were constructed with the

understandings defined by Frayer, Fredrick, and Klausmeier (1969):

       

   

44  

• concept instances are comprised of attributes which are relevant or irrelevant to the

concept

• non-examples lack one or more of these characteristics

• concepts have names and can be defined “structurally, semantically, operationally, or

axiomatically

• concept instances may be related to supraordinate, coordinate, and subordinate concepts

These considerations were used in the construction of the questions for the first, multiple-

choice assessment and also for the coding interpretation of the second, nonsense word

assessment. The first was not coded due to the fact that there was only one correct answer for

each prompt, the key of which is provided in Appendix L for both grade four and grade five. The

nonsense word tool went through first cycle and second cycle coding. There was a key for

assessing the clearly correct and clearly incorrect responses in the first cycle (Appendix M for

grades four and five). Second cycle coding was completed for further clarification of the student

responses that were “unclear” after the first cycle coding.

For both assessments, data was collected by individual class and also by the entire grade

levels (grade four and grade five). This was done by reviewing the number of correct vs.

incorrect responses on the multiple-choice assessment, and for the nonsense words, it was a

collection of the number of responses clearly correct, clearly incorrect, incomplete/missing, or

unclear (and therefore in need of further review in second cycle coding). Results were

represented in a chart as well as a bar graph for a more visual representation. Again, this was

completed for each of the two researcher created student assessments. There was an additional

chart and bar graph to represent the updated results after second cycle coding of the nonsense

words instrument.

       

   

45  

Analysis: Teacher focus groups and interviews. Teacher interviews were conducted

individually and in focus groups. The four focus groups were made up of grade 4 ELA teachers

and special educators, grade 4 science teachers and special educators, grade 5 ELA teachers and

special educators, and grade 5 science teachers and educators. Individual interviews were

conducted with the math/science Curriculum Leader and the English/social studies Curriculum

Leader. The questions were open-ended in nature so as to not bias the teacher responses in any

way (again, questions are located in Appendices H and I). Throughout the course of the study,

the teachers were not necessarily working together to provide instruction and administer

assessments. Each teacher was practicing within his or her own classroom. Therefore, the focus

group format will allow the “group members to respond to each other’s points, agreeing,

disagreeing, or modifying in any way they choose,” at the end of which, “the researcher will

encourage the group to come to a conclusion” about the students’ performances and use of

morphological understandings and transference (Rubin and Rubin, 2012).

Interviewer strategies were utilized to minimize focus group flaws. For example, all

responses were correlated with collected data where applicable to avoid the possibility that a

teacher or teachers may “make up answers” to avoid embarrassment or the perception that they

may be portrayed negatively for something done (or not done) in their class (Krueger and Casey,

2009). It was also be necessary for the interviewer to carefully “serve as a leveling force” in

moderating the focus group so that no “individuals c(ould) influence results” by dominating the

conversation (Krueger and Casey, 2009).

Responsive interviewing was utilized so that it would be possible to “elicit…examples,

narratives, histories, stories, and explanations, grounding answers firmly in the experiences of

the interviewees” (Rubin and Rubin, 2012) regarding student interaction with science terms in

       

   

46  

direct relation to using affixes to derive meaning. In this forum, teachers were able to share

historical information possibly not observed in the classroom. Utilizing responsive interviewing

allowed the research design to “evolve throughout the study” and for the interviewer to “hear

what (wa)s said and change direction to catch a wisp of insight, track down a new theme, or

refocus the broader questions” of the interview (Rubin and Rubin, 2012). When it was perceived

that an interview subject may not be comfortable sharing an answer in the focus group or

individual interview, an e-mail communication would have been used to follow-up so that the

interviewer may “respectfully listen to what (the subject) has to say” in a less public or direct

atmosphere (Rubin and Rubin, 2012). While this remained an option available for clarification,

its use did not become necessary.

There were two purposes for the questions that subjects were asked. First, the fidelity to

the vocabulary program could vary by teacher and team. In order to get an accurate

understanding of the influence or lack of influence that this approach had on the student’s

capacity to transfer their morphological understanding to a different content area, it first needed

to be clearly understood just how much of the program had been used and to what extent of

explicitness and mastery.

The first round of questions was to establish the type of vocabulary instruction that was

used during the school year, as well as to what extent it had been used. While these questions

were asked in the focus groups, they were repeated in the individual interviews. This was to

account for the possibility that individual teachers may have felt uncomfortable discussing their

level of fidelity to the school’s adopted program in front of their peers. It was also important to

determine whether the Curriculum Leaders’ expectations and understanding of what teachers

were doing in class was congruent with what the teachers reported.

       

   

47  

The next set of questions focused on the teachers’ observations of their students’ level of

vocabulary growth, achievement, difficulty, and challenges for both the aggregate of their classes

and the disaggregate of the student subgroups including students identified as having special

needs, low income, or levels of disengagement. It also included any examples that they could

provide of any transference of morphological skills during the year that they may have observed.

It was critical that the researcher was not influenced by “background and experiences,”

and therefore avoided “selective perception” in the analysis of participant responses (Krueger

and Casey, 2009). One way to avoid this was to use “verifiable analysis (Krueger and Casey).

The researcher utilized transcripts and provided oral verification of important takeaways.

Through this means, other reviewers should be able to come to the same understandings upon

review of the data. The Classic Approach of analysis was the strategy used to systematically

break down the information provided in the focus groups and the interviews. The main questions

were ordered with the responses of each group or individual. In this way, it was possible to

“identify themes and categorize results” (Krueger and Casey, 2009).

While some note taking was done during all focus groups and interviews, all that were

conducted in person were recorded and transcribed. All participants were asked for permission

before a recording device was used. The recording was done for the purposes of transcription and

was in the form of an MP3 file recorded on an iPhone 4 utilizing the Rev Voice Recorder: Audio

Transcription app. This file was sent to Rev Transcription Service. This company was selected

due to their 98% accuracy rate and commitment to confidentiality. The files were transmitted

utilizing 128-bit SSL encryption for security. The final transcription was in Microsoft Word

format.

       

   

48  

The interviewer and the subjects, to insure accuracy in language, context, and intention,

then reviewed the transcripts. Stakeholder checks “enhance the credibility of findings” (Thomas,

2006). Participants were provided with copies of their focus group and/or individual interview

transcripts and had the opportunity to “immediately correct errors of fact or challenge

interpretations” (Thomas, 2006)

Analysis: Coding. The researcher corrected both of the researcher-created student

assessment tools utilizing the answer keys. Only the second assessment, the nonsense word tool,

needed to be coded. The prompts were purposefully created as Production Items (Frayer,

Fredrick, Klausmeier, 1969). These allowed for the students to provide answers that did not rely

on any suggested prompts. Therefore, the answer key for first cycle coding was not able to

provide clarification on any unclear responses. Therefore, second cycle coding was needed. The

first cycle coding was Descriptive Coding as it “leads primarily to a categorized inventory,

tabular account, summary, or index of the data’s contents” (Saldaña, 2013). In order to be

accurately represented, great care was taken that the description was factual and contained rich

detail.

The second cycle coding was needed for discovery or clear understanding of the potential

complexities, so second cycle Pattern Coding was done for further “classifying, prioritizing,

integrating, synthesizing, abstracting, conceptualizing, and theory building” (Saldaña, 2013). All

of the passages that were coded with the same short noun phrases to indicate topics were taken

from the text and reclassified and reorganized based on a further analysis in order to make

meaning form the text. The Pattern Coding was used “as a stimulus to develop a statement that

describes a major theme, a pattern of action, a network of interrelationships, or a theoretical

construct from the data” (Saldaña, 2013).

       

   

49  

A professional transcription was completed at the conclusion of every focus group and

interview. These documents were read thoroughly, and an initial, Descriptive Coding was

completed. It was the intention that the codes indicate any findings, and that these findings

would “display multiple perspectives from individuals and be supported by diverse quotations

and specific evidence” (Creswell, 2009). As final interpretations, lessons, theories, or new

questions emerged, they were reported with rich descriptive detail (Creswell, 2009). Second

cycle pattern coding was used for the focus groups and interviews as well to determine themes in

their answers, anecdotes, and examples.

In the constructivist-interpretivist paradigm, different determinations may be made from

the same data. However, while there may be “multiple meanings of a phenomena in the minds of

people who experience it as well as multiple interpretations of the data” (Ponterotto, 2005), there

can be description that is detailed and rich enough to create a clear understanding of one

interpretation with evidence that is credible enough to substantiate the resulting theory and/or

consensus.

Trustworthiness  

I was conscious that, while I was not able to fully eliminate all biases, I must “take care

to identify and monitor” what they were throughout the course of the study (Merriam, 2009). It

was important that the experience of the participant vs. the researcher came through. Therefore,

the study needed to focus on the “emic” (universal) vs. the “etic” (unique to individual)

perspective.

Due to the researcher role as “participant” in this study, care was taken so that the

participants did not believe they sensed bias. Some of the teachers in the study were former

       

   

50  

colleagues, and it was important that they understood that there was no hypothesis or theory

being tested in this study; rather, it was the objective of the researcher to see if there was or was

not transference of using morphology as a strategy from one content area to another and whether

this increased student comprehension of concepts or not. Due to the researcher as participant

model, it was necessary to assure teachers that there was no vested interest in the results either

way so as not to skew the data collected.

There were specific strategies employed to increase the validity of the study. First, I

triangulated collected data by “converging several sources of data or perspectives from

participants” (Creswell 2009). This data included documents, interviews, and focus groups.

Second, follow-up interviews or “member checking” was utilized to guarantee the

accuracy of transcription and maintenance of the intentions of the participant responses. Also

referred to as “respondent validation” (Maxwell, 2005), this not only allowed me to be sure I did

not misinterpret a participant’s responses, but it also helped me to identify and address any biases

that I may have had as well.

Third, if there was an indication that my background as an English teacher would bring

bias to particular findings, this would have been acknowledged and described clearly. Also, if

any negative or contrary findings were revealed during the course of the study, they too would

have been shared and explained. In addition, if there were an opportunity for peer debriefing, this

would have been embraced and utilized.

Finally, a “stakeholder check” was completed (Thomas, 2006). All subjects were invited

to read, comment on, and/or correct the summary of the focus group/interview in which they

participated.

       

   

51  

Limitations  

Generalizability is not a typical feature of qualitative studies. However, there were

characteristics that may be considered generalizable in this study. First was the fact that “there is

no obvious reason not to believe that the results apply more generally” (Maxwell, 2005). In this

case, there were specific findings or themes that are certainly replicable to students of the same

age and grade in other school systems or even to students of different ages or grades within this

same school system.

A second consideration was that “the generalizability of qualitative studies is usually not

based on explicit sampling of some defined population to which the results can be extended, but

on the development of a theory that can be extended to other cases” (Maxwell, 2005). While the

students in this study were purposefully chosen to represent students who are now “reading to

learn” vs. “learning to read,” there may be different developmental ages and grades at which this

occurs elsewhere, thus allowing for more generalizability in that context.

Additionally, the communities in which this study took place have a largely

homogeneous student population. While the root of this study involves language, there are no

students currently identified in Billings Intermediate as Limited English Proficiency, and the

low-income rate, while present, is extremely low compared to the Massachusetts rate. This limits

the understanding of the transferability of this study to students in schools that are diverse and

include a higher Limited English Proficiency population. Since this study focused on language

acquisition, it may have specific implications of usefulness with these populations in particular,

so this is an area that should be investigated more fully.

       

   

52  

Chapter IV: Report of Research Findings The purpose of this study was to examine the potential effect that a morphological

approach to vocabulary acquisition may or may not have on the conceptual understanding of

scientific terminology. In addition, the extent to which a student may use the word attack

strategy of disaggregating known Greek and Latin word parts to identify the meaning of

unknown (or nonsense) content words in science was also an area of focus.

Research Questions  1. A. How do students apply their understanding of Greek and Latin word parts

in ELA to acquiring vocabulary and understanding new concepts in science?

B. What contributes to student application of morphological strategies learned in ELA to

their study of vocabulary in science?

2. A. To what extent do students utilize their understanding of morphemes to decode and

comprehend or conceptualize unfamiliar scientific vocabulary?

B. How do students hypothesize the meaning of nonsense scientific vocabulary

utilizing their understanding of Greek and Latin word parts?

A total of eighteen classroom teachers (nine from grade four and nine from grade five),

four special education teachers, and 397 students (201 in grade four and 196 in grade 5)

participated. The data collected centered around five main components. These consisted of one

teacher assessment and two student assessments created by the researcher as well as teacher

focus groups by grade level and content area and individual interviews with the Mathematics and

English Language Arts Curriculum Leaders.

       

   

53  

Teachers in this school work in partner teams. One teacher in each pair is responsible for

teaching math and science and the other teaches ELA and social studies. Therefore, teacher

“teams” report scores as each partner shares the same set of students. One grade four classroom

teacher and one grade five classroom teacher declined participation in the study. Both of these

teachers were ELA/social studies teachers. Because the student assessments were administered

during the students’ regularly scheduled science classes, the lack of this teacher participation had

no effect. Also, since they each have a “partner” teacher that shares the responsibility of teaching

the same students, this did not affect whether any of the classes could take part in the study or

not. The entire school is made up of grade four and grade five only, and every student was

invited to participate. A total of 31 students in grades 4 and 5 did not participate in the study

either through parent exclusion or self-exclusion. These students were not penalized in any way

and continued with regularly scheduled classwork. This resulted in a total sample size of 397

students taking part in the study (201 in grade four and 196 in grade five).

Teacher Assessment (Pre and Post Tests)   The vocabulary series that is being used in grades four through twelve is Growing Your

Vocabulary: Learning from Greek and Latin Roots (A and B) by Prestwick House (Moliken, Ed.,

2008). In it, there is an aligned scope and sequence of word parts, and there are a series of units

that teachers generally take about two weeks to complete. At the end of five units, there is a

cumulative, multiple-choice assessment. The teachers at Billings Intermediate administer this as

a summative measure at the end of the set of each five units (approximately ten weeks of

instruction). In addition, the ELA Curriculum Leader administered this assessment as a pre-test

as well as a post-test to document the achievement and growth of her students. She maintained

       

   

54  

this data this year for the two classes that she teaches with her partner colleague (n=44). At the

time of this study, the classes had completed two, five-unit cumulative assessments. She

presented this archival information for these classes for review without student names on them.

The results for the two classes (44 students):

• Pre-test #1: 53.3% to Post-test #1: 80.5%

• Pre-test #2: 62.3% to Post-test #2: 80.6%

While the post-test scores remained almost identical, the pre-test scores increased 9%

from the first pre-test to the second pre-test.

Student Assessment #1, Multiple-Choice Vocabulary Assessment   The first student assessment was designed to measure each child’s demonstration of

comprehension of vocabulary words containing Greek and Latin word parts that were part of

their grade level curricula. This included three types of vocabulary words: words that were part

of their English Language Arts (ELA) curriculum, words that were part of the science

curriculum, and words that they would have encountered in both the ELA and science curricula.

The Curriculum Leader for English Language Arts (ELA) and the Curriculum Leader for

Science provided the vocabulary words that they confirmed were the mandatory list of words in

each subject for grades four and five. All words chosen for Vocabulary Assessment #1 were

taken from these lists. Only words that contained a Greek or Latin root were selected. Each was

then identified as “ELA only,” “Science only,” or “Both ELA and Science,” depending on where

they were found in the two content area curricula.

This assessment was given to each of the five teachers in grade four and the five teachers

in grade five that are responsible for teaching math and science. After collecting student

       

   

55  

exemption forms from those students whose parents waived them from participation, the teachers

asked which students did not want to participate. These students were excluded from the study as

well. In addition, if a child began the assessment and at any point identified that they did not

want to continue, they were told they could stop, and they did so without penalty. Any “blank”

answers on the assessment were identified as such and did not count as a “correct” or “incorrect”

response. Most students, however, completed the full assessment.

The science teachers administered the Vocabulary Assessment #1 in their regularly

scheduled science classes. Each of these teachers repeated this administration with the students

in their “partner teacher’s” class. These are students that they are responsible for teaching math

and science to as well. These were collected by the classroom teachers and labeled by grade level

and teacher team.

The assessment prompts were designed specifically to measure students’ conceptual

understanding vs. their rote memory of the word and/or its definition. This was done by crafting

the questions in accordance with the research in “A Schema for Testing the Level of Concept

Mastery.” by Frayer, Fredrick, and Klausmeier (1969). Each question and set of answer choices

was created utilizing one of the models presented in this research to insure the most reliable

measure of conceptual understanding. These included attributes which are relevant or irrelevant

to the concept, non-examples lacking one or more of the conceptual characteristics, defining

concepts “structurally, semantically, operationally, or axiomatically,” and relating choices to

“supraordinate, coordinate, and subordinate concepts” (Frederick, Frayer, and Klausmeier,

1969).

       

   

56  

Every question on the assessment was created using these characteristics, and there were

an equal number of questions representing each as well. Figures 4.1 (grade four) and 4.2 (grade

five) provide examples of questions from the assessment that show each of these characteristics.

Figure 4.1. Examples of questions and the conceptual indicators they represent for grade four.

Grade 4

Attributes are relevant or irrelevant to the concept

A vertebrate is: A. An animal that has gills B. An animal that has no backbone C. An animal that has claws D. An animal that has a backbone E. An animal that has lungs

Non-examples lacking one or more of the conceptual characteristics

Which of the following is NOT an example of something thermal?

A. An ice cube B. The sun C. A fire D. A heated oven E. A lit match

Defining concepts “structurally, semantically, operationally, or axiomatically”

Dehydrate means: A. To lose or remove water B. To do something every year C. To be easily heard D. To measure something E. To move water in a container

Relating choices to “supraordinate, coordinate, and subordinate concepts”

Choose the sport that is aquatic: A. Baseball B. Swimming C. Soccer D. Football E. Basketball

       

   

57  

Figure 4.2. Examples of questions and the conceptual indicators they represent for grade five.

Grade 5

Attributes are relevant or irrelevant to the concept

Meat-eating animal usually have: A. Soft fur and tails B. Small teeth and tails C. Scales and fins D. Large teeth and claws E. Small eyes and no claws

Non-examples lacking one or more of the conceptual characteristics

Which of the following would NOT be eaten by an herbivore: A. Apple B. Hamburger C. Carrot D. Lettuce E. Peanuts

Defining concepts “structurally, semantically, operationally, or axiomatically”

The process in which plants take in light, carbon dioxide, and water to produce sugar and oxygen is:

A. Photosynthesis B. Chlorophyll C. Cell respiration D. Transpiration E. Adaptation

Relating choices to “supraordinate, coordinate, and subordinate concepts”

Which of the following is NOT a word that has to do with family or family members?

A. Maternal B. Paternal C. Thermal D. Matriarch E. Fraternal

Student answers were recorded and organized by question, subject (ELA, science, or

ELA and science), class, and grade. The grade four and grade five assessments both had twenty

questions. Each prompt provided five options from which students chose their answer. There

were three types of responses recorded: correct, incorrect, or no response (left blank). In all of

grade four, the number of correct responses overall ranged from 73.5% to 81.3%, representing a

standard deviation of 3.7%. In grade 5, the correct responses in all classes were between 78.9%

       

   

58  

and 84.7%, with a standard deviation of 2.2%. A complete distribution of the responses by

teacher team and grade is provided in Figures 4.3 and 4.4.

Figure 4.3. Totals of correct and incorrect responses, including standard deviation, for Student

Assessment #1 in each teaching team in grade four.

Grade 4 Totals – Student Assessment #1 by percent

Team Correct response

Incorrect response

No response Total responses (n)

Team 4A 81.32 18.68 0.00 760

Team 4B 79.88 19.63 0.49 820

Team 4C 75.00 25.00 0.00 780

Team 4D 73.49 25.35 1.16 860

Team 4E 73.50 25.13 1.37 800

AVERAGES 76.60 22.76

Standard deviation

3.70 3.31

       

   

59  

Figure 4.4. Totals of correct and incorrect responses, including standard deviation, for student

assessment #1 in each teaching team in grade five.

Grade 5 Totals – Student Assessment #1 by percent

Team Correct response

Incorrect response

No response Total responses (n)

Team 5A 78.90 21.10 0.00 740

Team 5B 83.10 16.90 0.00 720

Team 5C 84.70 14.30 1.00 860

Team 5D 81.67 17.90 0.43 840

Team 5E 83.00 16.20 0.80 760

AVERAGES 82.27 17.28

Standard deviation

2.17 2.51

Calculations for correct and incorrect responses were also completed by content area.

Each vocabulary word represented in the assessment was associated with the students’ ELA

curriculum or science curriculum. A third category was created for words that appeared in both

their ELA and science curricula. These totals are represented in Figure 4.5.

Figure 4.5. The percentage of correct responses for words taken from each content area curricula.

Grade 4 Grade 5

ELA 81% ELA 81%

Science 67% Science 83%

ELA and science 83% ELA and science 83%

       

   

60  

Student Assessment #2, Nonsense Word Completion

The researcher also created the second assessment administered to students in their

science classes. This instrument was designed to measure students’ use of their skills in decoding

unfamiliar words by presenting them with nonsense words. This eliminated the variable that a

child may have been exposed to the word outside of school and might know the definition for

that reason. Each of the nonsense words was a compound word that contained a word part of

Greek or Latin origin that the students had studied in ELA, science, or both their ELA and

science classes and a common word part with which students would have familiarity (such as

“house” or “pencil”).

This tool was intended to measure students’ use of their understanding of the Greek and

Latin word parts to comprehend the concepts presented in the nonsense word. This was also

intended to identify whether students were able to transfer their understanding of the word part(s)

to new and unfamiliar words in order to decipher the concept or meaning of the word. This was

achieved by asking students to write a definition of the word in addition to creating a picture of

the concept represented. Soliciting responses in this open-ended manner represented the

Production vs. Selection method of choosing answers that has been identified as the more

accurate method of identifying true conceptual understanding (Frederick, Frayer, and

Klausmeier, 1969).

Further research by Dorothy Frayer and her colleagues at the University of Wisconsin

supports the connection between a students’ representation of a word in multiple forms and their

understanding of the concept behind the word using another means of communicating that

understanding (Frederick, Frayer, and Klausmeier, 1969). This led to the development of the

Frayer Model. This is a graphic organizer that prompts students to define the vocabulary word in

       

   

61  

their own terms as well as list characteristics, examples, and non-examples of the word. Many

practitioners have expanded this to include pictorial representations of a student’s understanding

of the word, particularly if the child has difficulty with written expression.

While all of the words of Student Assessment #2, which was called “That’s Nonsense!”

were, in fact, nonsense words, they were compound words made up of a simple word that does

exist and a Greek or Latin root that also exists. Each root chosen was one that was part of the

children’s ELA curriculum, science curriculum, or word parts that they had been exposed to in

both their ELA and science curricula in the current school year.

The science teacher of each teacher-team in both grades four and five also administered

this assessment. It was important that this be given during the students’ regularly scheduled

science class. The research question attempted to assess to what extent children would use their

understanding of morphemes to identify the meaning of unknown words. Since many of the

morphemes they would encounter were taught to them in their ELA class, it was important to see

if they would transfer their understanding of the word part learned in ELA to their science class

and a science-related term.

Teachers were given the option of which day of the week to complete the assessments

within a one-week window. Again, parents were allowed to exempt students from participation,

and the children were given the opportunity to exclude themselves if they did not wish to

complete the assessment without penalty. Teachers were also given the instruction to allow any

student to stop at any time or to leave prompts blank if they were struggling, just as they had

with the first assessment. Again, most students completed the task. In addition, several teachers

anecdotally shared in the focus groups conducted later that the students had described the

Nonsense Word Completion assessment as “fun.”

       

   

62  

First cycle coding for nonsense word assessment. Due to the fact that the answers were

open-ended, two levels of coding were used to analyze the results. The first was simple

descriptive coding to determine one of four possible responses in order to categorize the results

(Saldaña, 2009). This level of coding primarily considered the student’s written description in an

attempt to match it with the key word or words identified in the answer key created by the

researcher prior to the administration of the assessment. Student pictures were also used in

instances when the written description may have been unclear but the picture was very clear (ie:

unmistakable picture of the sun for “solar”). The designations for each answer included:

1. A missing or incomplete response. This was issued when both response boxes (written

explanation and drawing prompt) were left blank or an instance where such little was

recorded in one or both boxes that it was indecipherable. The symbol for this was “O.”

2. A clearly correct answer. This designation was issued only when a descriptive response

identified the exact wording or nearly the exact wording as the researcher answer key.

This answer key was created by utilizing the Greek or Latin word part of the nonsense

word only. No credit was given or taken away for any other part of the word. The symbol

for this type of response was “✓.” The complete answer key was created using the

definitions provided by the curricula and teaching materials given to the researcher by the

Curriculum Leaders, and, therefore, the definitions used for teaching the students. The

key for the first round coding is located in Appendix M.

3. A clearly incorrect answer. This score was assigned when a student had a clear

misunderstanding of the prompt. Common misunderstood answers were compiled based

on potentially misleading cues in the nonsense words. (ie: answers regarding “x-rays” for

words with “exo” in them.) These cues were not based on Greek or Latin roots that

       

   

63  

students would have encountered in either their ELA or science classes. Therefore, these

responses were considered incorrect. In addition, answers that merely re-stated the Greek

or Latin word part with no further clarification in the description or the picture provided

was also considered incorrect (ie: Hydrocar = “a car that is hydro” or “a hydro car”). The

symbol for this was “X.”

4. An answer that required further consideration or was unclear. This type of student

response did not use the exact wording on the answer key. There was also no clear

misunderstanding of the meaning indicated. In these instances, the researcher determined

that these responses required a second level of coding, including a close review of the

student picture drawn for interpretation. The symbol for this was “?”.

Figure 4.6 illustrates the examples of each type of student response that resulted in one of

the three descriptive codes (clearly correct, clearly incorrect, or unclear) for grade four, and

Figure 4.7 shows the same information for grade five. The only other indicator would be “O” for

a blank response box.

       

   

64  

Figure 4.6: Sample student responses for grade four first cycle, descriptive coding.

Grade 4 sample answers

Nonsense word

Correct key word(s) (per answer key)

Student description response Code assigned

aquaskates water “Shoes that let me skate on the water.”

thermoshoes heat, temperature “Shoes that heat themselves.” ✓ invertebracow backbone “Cow can be inverted.” X

terrahouse earth, soil, or land “House full of terror.” X solarexplorer sun Someone who explores solar

panels.

?

astropants star Pants for an astronaut. ?

Figure 4.7: Sample student responses for grade five first cycle, descriptive coding.

Grade 5 sample answers

Nonsense word

Correct key word(s) (per answer key)

Student description response Code assigned

florabook flower “A book about flowers.” ✓ trifeet three “Someone with three feet.” ✓ visiphone see, sight “Phone with a Visa card in it.” X

matricake eats “A cake in three pieces.” X

unistar one, single “Only star in a galaxy.” ?

respihelmet breathing, air “Helmet with a gas mask.” ?

The results for the first cycle descriptive coding for grades four and five are represented

in the following Figures 4.8 - 4.11.

Figures 4.8 and 4.9: First cycle descriptive coding results for all of grade four for each question

of the assessment based on the four possible answers of unclear, clearly correct, clearly incorrect,

       

   

65  

incomplete/missing. For each question listed, the subject(s) in which it appears in the curricula is

listed. (The results for each individual grade four class can be found in Appendix N).

Figure 4.8 Grade four first cycle descriptive coding results by assessment question.

GRADE 4

Code/Totals

QUES/SUBJECT ? √ X O Total ques Ques 1 ELA 11 172 10 1 194 Ques 2 Science 31 20 122 21 194 Ques 3 Both 58 92 41 3 194 Ques 4 ELA 18 82 87 7 194 Ques 5 Science 12 92 79 11 194 Ques 6 Both 20 136 33 5 194 Ques 7 ELA 134 13 40 7 194 Ques 8 Science 48 5 128 13 194 Ques 9 Both 136 16 30 12 194 Ques 10 ELA 48 48 86 12 194 Ques 11 Science 98 63 28 5 194 Ques 12 Both 114 41 33 6 194 Ques 13 ELA 16 29 136 13 194 TOTALS 744 809 853 116 2406

       

   

66  

Figure 4.9 Grade four first cycle descriptive coding results by assessment question (bar).

Figures 4.10 and 4.11: First cycle descriptive coding results for all of grade five for each

question of the assessment based on the four possible answers of unclear, clearly correct, clearly

incorrect, incomplete/missing. For each question listed, the subject(s) in which it appears in the

curricula is listed. (The results for each individual grade five class can be found in Appendix O).

0  20  40  60  80  100  120  140  160  180  200  

?   √   X   O  

       

   

67  

Figure 4.10 Grade five first cycle descriptive coding results by assessment question.

GRADE 5 Code/Totals

? √ X O Total ques Ques 1 ELA 5 168 19 1 193 Ques 2 Science 72 66 48 7 193 Ques 3 Both 81 59 35 18 193 Ques 4 ELA 1 176 11 5 193 Ques 5 Science 54 128 8 3 193 Ques 6 Both 14 54 99 26 193 Ques 7 ELA 24 123 39 7 193 Ques 8 Science 77 69 33 14 193 Ques 9 Both 47 72 46 28 193 Ques 10 ELA 19 34 113 27 193 Ques 11 Science 37 31 117 8 193 Ques 12 Both 36 97 50 10 193 Ques 13 ELA 16 161 10 6 193 TOTALS 483 1238 628 160 2349

Figure 4.11 Grade five first cycle descriptive coding results by assessment question (bar).

0  

20  

40  

60  

80  

100  

120  

140  

160  

180  

200  

?   √   X   O  

       

   

68  

A qualitative analysis at the first level of coding of each question for grade four and grade

five is provided below. This analysis will only consider responses that were clearly correct and

clearly incorrect. The second cycle coding will address answers that required a further

consideration. Incomplete or blank boxes were simply marked as such. Next to each word

prompt will be the ratio of correct to incorrect responses (also available previously in Figure 4.8

for grade four and Figure 4.10 for grade five) followed by a description of the most common

answers for each. Again, the number of responses that required second level coding is not

included here.

Grade Four  

1. “aquaskates” (172/10): This nonsense word included the morpheme “aqua” which

was identified as a word part the students learned in ELA. It required the word “water” as part of

the answer. The vast majority of responses (89%*) were clearly correct, most varying only on

what the skate’s relationship was to water. For example, “roller skates that work on water,”

skates that help you under water,” skates made out of water,” skateboard with water in it,” skates

that float on water,” skates that can go in (or “under” or “on” or “through”) water were all

accepted as correct. There were a small number of incorrect responses that described skates as

the color aqua or aquamarine. One response described a “person laying on a box with skates on

both.”

*Any categories reflecting 50% or higher of the responses for a single prompt will be noted by

total and/or percentage in this first cycle description.

2. “exorain” (20/122): This nonsense word included the morpheme “exo” which was

identified as a word part the students learned in science. It required the word “outside” as part of

       

   

69  

the answer. Most students applied this word part incorrectly (63%). However, correct answers

included “the outside of a rain drop,” “rain outside,” “water on the outside of the drop,” “rain

with a hard shell on the outside,” or “rain with the middle on the outside.” Incorrect responses

mostly centered on the misunderstanding that it was related to an x-ray. These answers included

several repeats of “see through rain,” “rain you can see in,” or “invisible rain.” Another cluster

of misunderstandings had to do with viewing the root as “x” or as a negative or opposite such as

“no more rain,” “rain that does not hit the ground,” “rain going up,” “raining on the inside,” or

“rain that never ends.” Other responses considered incorrect included “an exercise machine that

rains,” “steaming hot (or “dry”) rain,” “something that makes rain,” “strong rain,” or “ something

wide.” There were a small number of responses that simply restated the prompt; “it is exo rain.”

This is also the prompt that most students chose to leave blank (21).

3. “moneymeter” (92/41): This nonsense word included the morpheme “meter” which

was identified as a word part the students learned in both ELA and science. It required the

word(s) “measure” (or ”how much” or “counts”) as part of the answer. Correct responses

included “something that counts your money,” “a machine that measures money,” “something

that tells how much money you have.” Incorrect answers included confusion with the meter that

is used at car spaces or an ATM. For example, “meter for your car,” “meter that you put money

into and take money out of,” “machine that throws out money,” or “money dispenser.” Other

incorrect responses were “money that fell from the sky,” “someone meeting money,” and

“money making machine.”

4. “hydrocar” (82/87): This nonsense word included the morpheme “hydro” which was

identified as a word part the students learned in ELA. It required the word “water” as part of the

answer. Correct responses provided many variations on the possible meaning, but they all

       

   

70  

included water. Examples are “car that is pushed by water,” “car that runs on water,” “car that

drives on (or “under”) water,” “car that gives water,” “car made out of water,” “car that needs

water,” “car that waters the grass when you drive over it,” and “waterproof car.” One

misconception may have come from the students hearing about “hydraulic power.” While this

could have been considered for further review, as there is a potential relationship, it was

concluded that this was more sophisticated than students at this age would understand. It was

therefore determined that the relationship was purely a matter of a word part sounding “familiar”

rather than any indication the children were aware of the nature of water in a “hydraulic” system.

None of the pictures indicated differently. Therefore, responses that referred to speed, power, or

lift were considered incorrect. These included “power/turbo car,” “car that can go 3,000 miles an

hour,” “car that goes up and down,” “fast car,” “car that flies,” and “car with flames.” Other

incorrect answers were “car with dog shapes,” “car with pictures on it,” and “car that drives by

itself.”

5. “invertebracow” (92/79): This nonsense word included the morpheme “invertebra”

which was identified as a word part the students learned in science. It required the words “no

backbone” as part of the answer. Almost all correct responses simply stated that “the cow has no

backbone” or their pictures depicted an action that indicated such. An example is “the cow is on

the ground because he has no backbone.” If students responded that the cow “has a backbone,”

this was considered incorrect. Even though they were able to recall the relationship to backbone,

it was clearly the opposite of the conceptual meaning. Several incorrect answers had to do with

the misconceptions of inversion or conversion. These included “the cow can be turned upside

down,” “an inverted cow,” “cow can turn inside out,” “cow that can change to other animals,”

“cow that can go on land and water,” or “ cow that can twist.” Another confusion included

       

   

71  

vibration such as “a cow that vibrates” or “cow that vibrates and can invert in the dirt to hide.”

Other less recognizable errors included “an insect cow,” “a cow that can move on its own,” “cow

with a horn,” “mad cow,” and “a machine that has one hook.” This prompt also provided the

most entertaining (though still incorrect) responses (and pictures) when considering they were

provided by 8 and 9 year old children. These included “a cow with purple milk,” “a cow with

lasers,” “a cow that can ride a skateboard,” one eyebrow that is really long,” “a cow with a butt

for a face and a face for a butt,” and “a cow that says ‘invertebracow’ when it goes to the

bathroom.”

6. “thermoshoes” (136/33): This nonsense word included the morpheme “thermo” which

was identified as a word part the students learned in both ELA and science. It required the word

“heat/warmth” or “temperature” as part of the answer. Seventy percent of the concepts were

accepted as correct, and these included “shoes that have a temperature,” “shoes that give off

heat,” “shoes that are warm,” “shoes that keep your feet warm,” and “shoes that can measure

temperature.” The response “toasty warm clothes” was also accepted. While it referred to clothes

vs. shoes, it captured the salient part of the answer, which was the referent to “thermo.” Incorrect

responses varied and did not indicate a clear relationship between them. These included “shoes

that are (or “have”) a thermos,” “shoes that resist the rain,” “shoes with springs,” “a star

shower,” “going fast like the wind,” “shoes that shoot out candy,” shoes that keep feet cold,” and

“shoes that glow in the dark.”

7. “astropants” (13/40): This nonsense word included the morpheme “astro” which was

identified as a word part the students learned in ELA. It required the word “star(s)” as part of the

answer. This prompt was one of four that had 50% or more responses that required second cycle

coding for answers because they were neither clearly correct or clearly incorrect, therefore

       

   

72  

necessitating further review (134 or 69% of responses). Correct answers were “stars with pants,”

“star pants,” “pants made out of stars,” and “pants with stars on them.” Incorrect responses

included “pants that fill up by pulling a cord,” “pants that go on by themselves,” “warm fuzzy

pants,” “pants that get bigger,” “people go on earth,” and “pants that fit.” There were no

immediately recognizable relationships between these answers. Again, over one hundred more

responses needed further analysis.

8. “millistars” (5/128): This nonsense word included the morpheme “milli” which was

identified as a word part the students learned in science. It required the word “thousands” as part

of the answer. There were very few clearly correct answers (5). These were repeats of “1,000

stars” or “1/1000 of a star.” The incorrect answers were mainly the confusion with “million”

such as “a million stars” or “millions of stars.” Other incorrect quantities such as “ten stars,”

“100 stars,” or “many stars” were also not considered correct. The second common conceptual

misunderstanding was regarding size, specifically a small size. While an argument could be

made that this may come from an understanding of “milli” as 1/1000, it was determined, even

with a further analysis in second level coding, that there would not be enough information

provided in this assessment to determine that it was truly an understanding of “thousands.”

Therefore, answers such as “tiny, “small,” “mini” stars as well as “tiny, faded stars” were

considered incorrect. Finally, other, more unrelated (to one another) incorrect answers included

“stars in the military,” “stars that shine bright,” “stars that come in the daytime,” “stars in the

Milky Way,” “stars up in the sky,” “stars all named ‘milli’,” “people on a TV show,” and “a star

that you draw.” The total number of incorrect responses was 66%.

9. symmetrishapes” (16/30): This nonsense word included the morpheme “symmetri”

which was identified as a word part the students learned in both ELA and science. It required the

       

   

73  

word “same” or “similar” as part of the answer. This prompt had the most responses that

required further analysis (136). In other words, seventy percent of the students in grade 4

provided an answer that was not clearly correct or incorrect (based on the identified key words in

the answer key) and potentially provided more information in their drawing or contextually that

would give a better indication of whether it should be regarded as correct or incorrect. Again,

further description of this will be provided in the second level coding for this item. Clearly

correct responses included “shapes that have sides that are the same,” “same shapes but different

colors,” and “shapes that are exactly alike” (or “similar”). Clearly incorrect answers included

“not same shapes,” “weird shapes,” “stars you can trace,” and “shapes that are 3D.” Other

answers referenced “symmetry,” but did so incorrectly such as “shapes that never have a line of

symmetry” or “shapes with lines of symmetry” (accompanied by a drawing of an asymmetrical

shape.

10. “lunafood” (48/86): This nonsense word included the morpheme “luna” which was

identified as a word part the students learned in ELA. It required the word “moon” as part of the

answer. There were many variations of the definition, but that were all clearly correct in terms of

the morpheme. These included “food that is little moons,” “food on (or “from”) the moon,” “get

food from there and bring it back” (accompanied with a drawing of a crescent moon), “food

shaped like the moon,” “food for the moon,” and “food that grows on the moon.” There were

some clearly incorrect answers that came from a misunderstanding and identified the root as

“tuna” vs. “luna.” Far more incorrect answers were varied and seemingly unrelated to one

another. These were “food that people eat,” “food that makes you feel smart,” “food from a

store,” “small portions of food,” “plastic food used for pranks,” “sun made of food,” “bat food,”

“food that glows,” “when you eat breakfast for lunch,” “smelly food from a week ago,” “food

       

   

74  

that’s not good,” “small food,” “food made by Luna (person),” “the carpet company that sells

food,” “stars that are food,” and “food that cats and dogs eat.” A small number of students also

incorrectly answered “food that makes you crazy” or “food that makes you loony.” Again, while

there may be some understanding of the root here (from “lunatic”), it was determined to be too

much of a leap in the analysis to conclude it may be related to the concept of “luna,” and the

remaining evidence to review would not be enough to make such a conclusion. Therefore, these

responses were also considered incorrect.

11. “amphibicat” (63/28): This nonsense word included the morpheme “amphibi” which

was identified as a word part the students learned in science. It required the words “both” or

“water and land” as part of the answer. Half of student answers to this prompt required second

cycle coding (98). Most of the remaining responses were correct. These included many

repetitions of variations such as “cat that eats on land and water” and that “can breathe on land

and swim in the water.” Answers that were considered clearly incorrect were “cat that goes

wild,” “an animal,” “big construction machine,” “cat with weird features,” “cat with big ears,” “a

small cat,” “a cat on the rug,” “cat that can sing the ABCs,” “cat that’s invincible,” and “cat that

eats plants.”

12. “solarexplorer” (41/33): This nonsense word included the morpheme “solar” which

was identified as a word part the students learned in both ELA and science. It required the word

“sun” as part of the answer. Again, this was one of the four prompts providing the most need for

a second cycle coding as 59% of the answers needed further review. There were not many

variations on the clearly correct answers. They varied only as much as “sun explorer,” “person

who explores the sun,” and “only person to touch the sun and not die.” The clearly incorrect

       

   

75  

answers included “an expert explorer,” “explorer on the moon,” solar that explores,” “people that

explore stuff,” and “someone who explores space” (combined with a drawing of the moon).

13. “terrahouse” (29/136): This nonsense word included the morpheme “terra” which

was identified as a word part the students learned in ELA. It required the word “earth,” or

“soil/land/dirt” as part of the answer. Correct answers included “a house under the earth,” a

house in terrain” (with a drawing labeled with “grass” and “dirt”), “house in the ground,” and “a

house made of mud and dirt.” In terms of incorrect answers, this was the prompt that reflected

the largest number of clearly incorrect responses (70%). The vast majority of these came from

the misconception that “terra” referred to “terror” (one that the researcher concludes may

distinctively be a “Boston-area” error). Examples of this varied a great deal, but all represented

the same idea. These include “a house that tortures people,” “a house of terra” (accompanied by

a picture of a house with a thought bubble that said “Boo!”), “a haunted house,” and many other

varieties of “terrifying/spooky/creepy/scary house.” Some other, unrelated answers that were

also incorrect included “a humongous house,” “a house that grows each month,” “a house that

looks like a crown,” “a flying house,” “a house in the air,” “a very large house,” “a house in

space,” and “a house so big that a Terranasaurus (sic) can fit in it.”

Grade Five  

1. “florabook” (168/19): This nonsense word included the morpheme “flora” which was

identified as a word part the students learned in ELA. It required the word “flower” or “plants”

as part of the answer. Most students answered this prompt similarly and correctly (87%). The

most common responses were “a book about flowers” (or “plants”) and “a book that has flowers

in it.” Other students identified the word as “a book shaped like a flower,” “flowers with books

       

   

76  

for petals,” “and “a book on plants to help with research.” A few children misidentified the

prefix and answered, “in Florida reading a book,” “a floor with a book on it” (or “under” it), or

“step on a very hard book.” These answers were clearly incorrect. Other responses that were

wrong included “a nice made book,” “me reading a book,” “an open book,” and “a dictionary.”

2. “biohouse” (66/48): This nonsense word included the morpheme “bio” which was

identified as a word part the students learned in science. It required the word “life” as part of the

answer. Responses that were clearly correct each identified a house differently, but they all

correctly referred to the key words in identifying the concept in question. Answers included

variations such as “a house of living organisms,” “house full of life,” “house with living things,”

“where one lives,” “house that looks alive,” “house that is a living thing,” and “house that is

made of living things.” This prompt required the second coding review of several student

answers (72) because of all the references to words containing “bio.” The pictures will need to be

analyzed for further information. However, there were some answers that were clearly incorrect.

These included those that showed student misconceptions such as “two houses” or “two houses

that look the same.” This is a result of the confusion with “bi,” meaning “two.” One answer that

several students shared was “the study of a house” or “to study a house.” The children utilized

the “ology” part of the word “biology” for this misconception. Other answers included, “a foto of

a house,” “buying a house,” “a house of maps,” “solar panels on a roof,” “a cold house,” “a

house of famous people,” “a big house,” and “house of one type of person.”

3. “omnicolors” (59/35): This nonsense word included the morpheme “omni” which was

identified as a word part the students learned in both ELA and science. It required the word “all”

(or “many/multiple”) as part of the answer. Correct answers included a clear indication that there

was “many” of something or referred to “all” of something. The most common answers were

       

   

77  

“many colors,” “all different colors,” “all of the colors,” or “color that’s made up of many

colors.” However, many students made a clear error in their attempt to apply this morpheme.

Clearly they remembered the words “omnivore” and “herbivore” as each of them attempted to

define the word as something that “eats.” This was confusion between “omni” and “vore,”

meaning “one who eats.” Answers that reflected this misunderstanding were “a color that eats

plants and animals,” “meat colored,” “colors that look like food,” “paint that is alive and likes to

eat things,” “colors for dinosaurs,” and “colors that eat meat.”

4. “trifeet” (176/11): This nonsense word included the morpheme “tri” which was

identified as a word part the students learned in ELA. It required the word “three” as part of the

answer. The majority of students defined this nonsense word with a correct representation of the

morpheme (91%). The key word was “three,” but students offered many creative variations that

were all correct. These included “three feet,” “three legs and three feet,” “a monster with three

feet,” “three feet together,” “three-toed feet,” and “shoes for people with three feet.” Still others

referred to the measurement as with “one yard, or three feet” and “a yardstick that is three feet

long.” Answers that were not accepted as correct did not specify “three” such as “a tree feet,”

“my brother has tiny feet,” “two feet,” “tires with feet,” and “a lot of feet” (with a picture

representing five feet).

5. “electrodog” (128/8): This nonsense word included the morpheme “electro” which

was identified as a word part the students learned in science. It required the word(s) “electric” or

“electricity” as part of the answer. Such responses included “dog that has electrical energy,”

“dog that gives off electricity,” “dog getting zapped by electric fence,” “it runs on electricity,”

“dog creates electricity,” “dog with electric touch,” “dog that uses electricity,” “dog with electric

powers,” “dog with electricity coursing through its system,” and “dog created by electricity.”

       

   

78  

There were only eight students whose answers were clearly incorrect. These included the most

common answer, which was “fast dog,” and one that was “dog that goes up and down non-stop

for an hour.”

6. “respihelmet” (54/99): This nonsense word included the morpheme “respi” which was

identified as a word part the students learned in both ELA and science. It required the word(s)

“breathe” or “air” as part of the answer. As seen in other prompts, the children had many

different ideas of what this nonsense word represented; however, they all clearly indicated the

correct concept of the morpheme. Examples were “a helmet with holes in it to get air in,”

“helmet I can breathe in,” “helmet that can breathe oxygen,” “helmet that lets you breathe

underwater,” and “helmet with air coming out.” Over half of the answers given were clearly

incorrect (51%). Most of these did not have a clear connection or indication of what the

misconception may have been. These include “a helmet that helps if you are hurt,” “a phone

helmet,” “helmet with a defibrillator attached,” “helmet with more padding,” “helmet that

expired,” “helmet inside the body,” “helmet that has respect,” “iron helmet,” and “snake helmet.”

Other clearly incorrect answers described the helmet as “old and dirty,” “new,” “reassembled,”

“protective,” “foam,” and “spiked.” Three students had answers that resembled a possible

confusion with “perspiration.” These were “a sweating helmet,” “a helmet that soaks up sweat,”

and “absorbs sweat.” Two other students included “recipe” in their answers, thus indicating

another misunderstanding when they pronounced the morpheme of this “word.”

7. “unistar” (123/39): This nonsense word included the morpheme “uni” which was

identified as a word part the students learned in ELA. It required the word(s) “one” or “single” as

part of the answer. Most students provided clearly correct responses for this prompt (64%).

There were many answers that were similar such as “one star,” “one whole star,” “one star made

       

   

79  

up of many stars,” and “one star in the sky.” There were others that demonstrated understanding

of the concept as well, but worded their descriptions a little differently. These included “a one

pointed star,” “star with one dot,” “stars that are together as one,” “one star that’s different,”

“someone looking at one star,” and “more than three stars coming together as one whole.”

Students who produced clearly incorrect answers seemed to understand that the concept involved

quantity, but they provided the wrong amount with answers such as “three stars or more,”

“double stars,” “star with three sides,” “group of stars,” and “lots of stars.” The remainder of the

answers that were incorrect were not seemingly related to one another. These were “a type of

star,” “star with units,” “a rockin rock star,” “an underground star,” “small star,” “stars that are

getting bigger,” and “stars that are singing.”

8. “visiphone” (69/33): This nonsense word included the morpheme “visi” which was

identified as a word part the students learned in science. It required the word “see” (or “sight” or

“look”) as part of the answer. The most common clearly correct answer was a variant of “phone

that you can see other people when you call them” or “see the people while you talk to them.”

Other answers described the phone itself such as “phone with a clear case so you can see what’s

inside,” “phone with really big numbers so everyone can see it,” “seeable phone,” and “see

through phone.” Finally, other answers that were correct included what the phone could do,

including “phone that makes you see invisible things,” “phone that helps you see,” and “phone

that sees.” While the definitions varied, they were all correct because they contained the key

word(s) indicating conceptual understanding of “visi.” Among the clearly incorrect answers were

“a news phone,” “a phone you can visit places with,” “phone with a visor,” “new brand of

phone,” “phone you can’t see,” and “a photo of a phone.” A small number of students’ answers

       

   

80  

incorrectly made reference to the credit card. These included “a phone with a Visa card on it” or

“a card phone.”

9. “chocolovore” (72/46): This nonsense word included the morpheme “vore” which was

identified as a word part the students learned in both ELA and science. It required the words

“one that eats” (or “feeds”) as part of the answer. All of the correct answers were variations of

something eating chocolate. Such definitions included “a living thing that only eats chocolate,”

“chocolate that only eats chocolate,” “chocolate eater,” and “dinosaur that only eats chocolate

bars.” Answers that were clearly incorrect were “a new chocolate brand,” “chocolate that speaks

a different language,” “chocolate lava,” “a cereal,” “a chocolate flower,” “chocolate animals,” “a

variety of chocolates,” and “a new flavor of chocolate.” Only a couple of students confused the

root as used in other words and answered, “a chocolate omnivore” and “chocolate shaped like a

carnivore,” and these were also incorrect.

10. “matricake” (34/113): This nonsense word included the morpheme “matri” which

was identified as a word part the students learned in ELA. It required the word “mother” as part

of the answer. While all of the clearly correct answers supplied by students incorporated the

word “mother,” there were many different phrasings. Among them were “a mother who eats

cake,” “a birthday cake for mom,” “momma cake,” “cake made of your mom,” “cake in the

shape of a mom,” “cake only made for mothers,” “a motherly cake,” “my mom made cakes,” and

“a cake that is the mother of two cupcakes.” However, most students (59%) answered this

prompt incorrectly. The more common misconceptions were those by students who confused

“matri” with meaning “mattress,” “mat,” “wedding” (matrimony), and “three” (tri). This was

evident in definitions such as “mattress cake,” “three layer cake,” “three cakes,” and “wedding

cake.” Other students had many different interpretations and provided incorrect answers. In fact,

       

   

81  

this prompt provided more varied answers that were incorrect and seemingly unrelated to any

other. These responses included “cake that tells time,” “cake in a lot of different pieces,” “cake

that is happy,” “symmetrical cake,” “mature cake,” “maze made out of cake,” “soft cake,” “cake

that plugs in,” “math cake,” “metric cake,” “complicated cake,” “crazy cake,” “cake that is first

to eat,” “really big cake,” “elephant cake,” “robot cake,” “money cakes,” “king and queen

cakes,” “father cake,” and “an oven that makes cakes.”

11. “photofood” (31/117): This nonsense word included the morpheme “photo” which

was identified as a word part the students learned in science. It required the word “light” as part

of the answer. While most students incorrectly identified this morpheme, the ones that were

correct did focus on the definition of “light.” There were many variations of the definition

considering the addition of “food,” but they were all clearly correct. These included “food that

absorbs light,” “food that grows from sunlight,” “light is food for plants,” “light food,” “the light

is in my food,” “a bologna sandwich made of light,” “food you make using light,” “a food for

light,” “food with light on it,” “an apple that likes sunlight,” “and “an apple giving off light.”

Again, most students (61%) misidentified this morpheme’s concept. Of the clearly incorrect

answers, all but one included the words “photograph,” “picture,” or “camera,” but gave no

indication of the understanding of “light.” Such answers included “food that can take a picture,”

“food and a camera,” “I took a photo of my food,” “camera made out of food,” “food you can eat

by looking at it in a photograph,” and “app to take pictures of food.” The only other clearly

incorrect answer was “changing food.”

12. “subdirt” (97/50): This nonsense word included the morpheme “sub” which was

identified as a word part the students learned in both ELA and science. It required the word

“under” (or “below”) as part of the answer. Half of the students (50%) correctly identified this

       

   

82  

morpheme’s concept in first cycle coding. Those responses that were accepted as clearly correct

included “something under the dirt,” “a sub under dirt,” “the layer of dirt under the first layer,”

“dirt that’s below the surface,” “dirt below water,” “underneath dirt,” and “an x-ray drawing of

dirt under topsoil.” There were also clearly incorrect answers. There was some indication that

students were making reference to the word “substitute” in answers such as “fake dirt,” and back

up dirt.” However, these did not indicate a clear understanding of the concept of “sub.” Other

clearly incorrect answers included “holographic dirt,” “dirt going back into the hole,” “subway

dirt,” “dirt that looks like from ages ago,” “twice as much dirt,” “person that’s a little bit dirty,”

“a sub full of dirt” (with a picture of a “submarine” sandwich roll), “extra dirt,” “a dirt

submarine,” and “dirt above the dirt’s crust.

13. “duoflag” (161/10): This nonsense word included the morpheme “duo” which was

identified as a word part the students learned in ELA. It required the word “two” or “double” as

part of the answer. Most students responded to this prompt by correctly identifying the concept

(83%). The clearly correct answers all contained the word “two,” but with differing definitions

when combined with into the compound “duoflag.” These correct answers were many repeats of

“picture of two flags,” “two flags connected,” “a flag with two countries,” “two flags on one

stick” (or “pole”), “a flag with two sides,” “two flags having a fight,” “two flags put into one,”

and “two flags stuck together.” The few clearly incorrect answers included “six flags,” “a game

like capture the flag,” “I had a duel to capture the flag,” “re-do flag,” “flag of dough” (with a

picture depicting a flag and the word “Pillsbury” written on it), and “a flag that makes pimples

go away.”

Second cycle coding for nonsense word assessment. The second cycle coding consisted

of pattern coding. This was done to further “classify” and “synthesize” student responses

       

   

83  

(Saldana, 2009). The second cycle coding was only used for those responses identified as

“unclear” in the first cycle coding to determine whether they would ultimately be classified as

correct or incorrect responses. However, in addition to further determining whether the unclear

responses were “correct” or “incorrect,” some of the patterns were also used to gain further

insight into the levels of student conceptual understanding.

Some student responses were more straightforward than others. Examples below show

answers provided by students that were all identified as “unclear” in the first round coding. The

second round coding took into account the student’s drawing as well as the context of their

descriptive sentence.

As the research of Frederick, Frayer, and Klausmeier indicated, a student’s level of

conceptual understanding can be measured most accurately with assessments that provided

information other than a simple definition of the vocabulary word, such as with the Frayer Model

(1969). Their research posits that student understanding can best be measured through other

prompts. Again, these include attributes relevant to the concept, attributes irrelevant to the

concept, non-examples, names that can be defined structurally, semantically, operationally, or

axiomatically, and concepts related to the supraordinate, coordinate, and/or subordinate

(Frederick, Frayer, and Klausmeier, 1969).

Therefore, any student response that was identified as “unclear” in the first cycle coding

was then analyzed for one of these indicators. Any that could be categorized as falling into one

of these ranges was accepted as a “correct” answer. Those that were not indicators listed or that

remained “unclear” after analysis were then considered “incorrect” responses.

       

   

84  

With this second round of coding complete, the student data results were recalculated to

include only correct, incorrect, and incomplete/missing responses. The results for the second

cycle pattern coding are represented in the following Figures 4.12 - 4.15.

Figures 4.12 and 4.13: Second cycle pattern coding results for all of grade four for each

question of the nonsense word assessment based on the three possible answers of correct,

incorrect, or incomplete/missing. For each question listed, the subject(s) in which it appears in

the curricula is listed. (The results for each individual grade four class can be found in Appendix

P).

Figure 4.12 Grade four, second cycle pattern coding results by assessment question.

GRADE 4 Code/totals

QUES/SUBJECT √ X O Total ques Ques 1 ELA 180 13 1 194 Ques 2 Science 35 138 21 194 Ques 3 Both 100 91 3 194 Ques 4 ELA 88 99 7 194 Ques 5 Science 99 86 11 196 Ques 6 Both 140 49 5 194 Ques 7 ELA 39 148 7 194 Ques 8 Science 6 175 13 194 Ques 9 Both 132 50 12 194 Ques 10 ELA 58 124 12 194 Ques 11 Science 136 53 5 194 Ques 12 Both 80 108 6 194 Ques 13 ELA 40 141 13 194 TOTALS 1133 1275 116 2408

       

   

85  

Figure 4.13 Grade four, second cycle pattern coding results by assessment question (bar).

Figures 4.14 and 4.15: Second cycle pattern coding results for all of grade five for each question

of the nonsense word assessment based on the three possible answers of correct, incorrect, or

incomplete/missing. For each question listed, the subject(s) in which it appears in the curricula is

listed. (The results for each individual grade five class can be found in Appendix Q).

0  

20  

40  

60  

80  

100  

120  

140  

160  

180  

200  

√   x   o  

       

   

86  

Figure 4.14 Grade five, second cycle pattern coding results by assessment question.

GRADE 5 Code/totals

QUES/SUBJECT √ X O Total ques Ques 1 ELA 170 22 1 193 Ques 2 Science 106 80 7 193

Ques 3 Both 116 59 18 193 Ques 4 ELA 177 11 5 193 Ques 5 Science 161 29 3 193

Ques 6 Both 59 108 26 193 Ques 7 ELA 141 45 7 193 Ques 8 Science 100 79 14 193 Ques 9 Both 80 85 28 193

Ques 10 ELA 45 121 27 193 Ques 11 Science 49 136 8 193 Ques 12 Both 129 54 10 193

Ques 13 ELA 175 12 6 193 TOTALS 1508 841 160 2509

Figure 4.15 Grade five, second cycle pattern coding results by assessment question (bar).

0  

20  

40  

60  

80  

100  

120  

140  

160  

180  

200  

√   x   o  

       

   

87  

Figure 4.16 Totals of correct and incorrect responses, including standard deviation, for the nonsense word assessment in each teaching team in grade four.

Grade 4 Totals – Nonsense Word Assessment by percent

Team Correct response

Incorrect response

No response Total responses (n)

Team 4A 45 50 5 37

Team 4B 49 43 8 39

Team 4C 39 60 1 40

Team 4D 48 45 7 40

Team 4E 43 54 3 38

AVERAGE 44.8 50.4 4.8

Standard deviation

4.02 6.87 2.86

Figure 4.17. Totals of correct and incorrect responses, including standard deviation, for the

nonsense word assessment in each teaching team in grade five.

Grade 5 Totals – Nonsense Word Assessment by percent

Team Correct response

Incorrect response

No response Total responses (n)

Team 5A 55 40 5 37

Team 5B 59 18 23 34

Team 5C 61 37 2 43

Team 5D 61 35 4 41

Team 5E 64.5 35 .5 38

AVERAGES 60.1 33 6.9

Standard deviation

3.47 8.63 9.16

       

   

88  

Figure 4.18. The percentage of correct responses for words taken from each content area curricula.

Grade 4 Grade 5

ELA 42% ELA 73%

Science 35.6% Science 54%

ELA and science 58% ELA and science 50%

A qualitative analysis at the second level of coding for each question for grade 4 and

grade 5 is provided below. After the nonsense word is listed, the number of responses that were

determined to be unclear in the first level coding, and therefore part of the second level coding, is

noted. The resulting number of correct vs. incorrect responses is then listed. Finally, a brief

explanation for that determination is given.

Each grade is followed with a chart that represents the amended results for the nonsense

word assessment with the updated number of correct and incorrect responses.

Grade Four  1. “aquaskates” (11 reviewed; 8 correct and 3 incorrect)

Answers that were considered correct included those that provided relevant attributes

such as “skates with fins” and “boots with boat propellers.” Students also defined the concepts

operationally (Frederick, Frayer, and Klausmeier, 1969), such as “skates that row” or “skates that

swim.” Another answer was given by relating the concept of “water” to a supraordinate concept

(Frederick, Frayer, and Klausmeier, 1969) by writing “skate in the ocean.” Further clarification

in some instances was provided by a review of the drawing provided. “Jet powered skates” had a

picture of skates on water waves and another response that was lacking any written description

was also a picture of skates on water waves.

       

   

89  

Responses that were deemed incorrect were “skates on liquids,” “skates in po??”

(indistinguishable last two letters), and “he is skiing.” A review of the drawings could provide no

evidence that the child understood the core concept to be “water.”

2. “exorain” (31 reviewed; 15 correct and 16 incorrect)

The correct responses mainly referred to “exoskeletons.” However, they provided some

indication, through their words or drawings, that the skeleton was on the “outside.” This was

important to note so that it could be distinguished from an incorrect response that merely

repeated a word with which they had familiarity since the exoskeleton was the context in which

they studied “exo” in science class. Therefore, if the student answered “an exoskeleton falling

form the sky,” “raining exoskeletons of animals,” or “raining with exoskeletons,” the student had

to provide a picture that showed bones or a skeleton on the “outside” of a drop of rain. Other

correct answers included rain “going around something” which was very similar in the picture to

going “outside” an object as well as other answers with wording as close such as “rain that was

developed from outer space” and “inside out rain.” Finally, phrases that provided relevant

attributes of the concept (Frederick, Frayer, and Klausmeier, 1969) were accepted as correct

answers such as “rain that is rock hard” and “hard rain from outer space.” One answer provided

an “outside” example of defining a concept structurally (Frederick, Frayer, and Klausmeier,

1969) by stating “the rain has an exoarm” accompanied by a drawing with a raindrop that had an

arm on the “outside” of it.

Incorrect responses included several that seemed to follow the misconception that “exo”

related to “xray.” This was seen in “rain you can see its exoskeleton,” “rain that’s hollow” and

“rain that allows you to see your bones.” Answers that indicated the children were confusing the

morpheme “exo” with bones or a skeleton only due to their use of the word “exoskeleton” were

       

   

90  

marked as incorrect. These included “rain with bones” bones coming down,” and “it’s raining

bones.” Other incorrect answers included “she’s in the rain,” “it goes on the horse and sha??s

(indistinguishable) away,” “rain from the inner core,” and “raindrops exhaling air.” There was no

evidence in the drawings or contexts that provided further information that these might be correct

in the second review.

3. “moneymeter” (58 reviewed; 8 correct and 50 incorrect)

The correct responses were those that indicated some form of understanding of

“measurement.” This was made clear in the second coding primarily through the students’

drawings. Examples included many several that stated “a meter that shows money” with a

drawing with notations on it such as “$500,” “$2.50 line” or “you have $5.00 left.” Others had

pictures only, but the pictures indicated a line that showed a growth of money on progressive

lines. Other answers included “a money ruler” and “it tells you your budget.”

Most of the incorrect answers were determined as such because they made particular

reference to the metric “meter” of three feet. This indicated once again that children might have

been noting a word that is familiar to them vs. applying the actual meaning of the root. Examples

were “money that tells size in meters,” “money that equals one meter,” “a meter with money on

it,” “a meter stick made of money,” “money in meters” and “money that is a meter long.” Other

incorrect answers were “holds money,” “the temperature of money,” and “thermometer with

money in it.”

4. “hydrocar” (18 reviewed; 6 correct and 12 incorrect)

The correct answers either provided clarification through their drawings or one of the

indicators of conceptual understanding. Examples of answers that were paired with pictures

depicting water were “keeps car moist,” “can drive on anything,” and “car that hydrates a

       

   

91  

person.” Other responses showed conceptual understanding through the indicators. This included

the use of a structural definition with “car that runs on hydrogen and oxygen,” relating the

concept to a coordinating concept with “a rain car,” and noting attributes that are relevant to the

concept with “always wet and keeps you hydrated” (Frederick, Frayer, and Klausmeier, 1969).

All of the incorrect answers were virtually the same in that they stated that the “car runs

on hydrogen,” “is made of hydrogen,” or was simply “a hydrogen car.”

5. “invertebracow” (12 reviewed; 7 correct and 5 incorrect)

During the second cycle coding for this prompt, two of the answers considered correct

upon further review were accepted because they provided examples of relevant attributes to the

concept (Frederick, Frayer, and Klausmeier, 1969). One student response was “a cow that is

paralyzed” and the other wrote “a flexible cow” with a picture of a cow that resembled a snake.

Every other answer that was changed from unclear to correct was due to the drawing that the

child provided. For example, when several students answered, “no vertebra,” the cow is an

invertebrate,” “or “cow is without a vertebra,” they provided pictures that depicted cows with a

back that was wavy or dipped in an inverted fashion so as to depict a missing backbone.

The responses that were considered incorrect did not have pictures that were helpful in

determining their conceptual understanding. The phrases they provided were “a messed up cow,”

“a cow that can twist,” “no vertebra,” “no bones,” or “invertebra cow.” While some of these

answers are similar or even identical to others that were accepted as correct, there was no further

evidence from their picture that provided a reasonable referent. They either depicted a “normal”

looking cow, one with lines above it, or simply one standing up.

6. “thermoshoes” (20 reviewed; 4 correct and 16 incorrect)

       

   

92  

One student provided a picture that offered a bit more insight into their understanding.

This student’s response was simply “has a thermostat.” However, their picture noted a meter on

the shoe they drew that had increasing temperatures noted on the side of it. Other students

provided information through the conceptual indicators. One example was a student who chose

an attribute relevant to the concept by stating concept, writing that they are “shoes that have

blankets.” Another child defined the concept operationally by writing “shoe that keeps out the

cold.” A third student provided a coordinating concept with “shoes that shoot fire” (Frederick,

Frayer, and Klausmeier, 1969).

Like other prompts, several of the student answers, upon further review, did not provide

additional information to indicate their understanding of the concept. Most of these answers

included that the shoe had “a thermostat” or “a thermometer” or even that it was a shoe “made

out of thermometers.” While these items are obviously related to the concept, neither the words

nor the pictures that went with them provided any indication that the children knew that the idea

was “heat” or “temperature.” Still other students responded that the shoe had a “thermos” in it,

could “tell the weather,” or that they were “shoes not found with thermal vision.” Again, there

was not enough evidence to determine that these were correct responses in terms of conceptual

understanding.

7. “astropants” (134 reviewed; 26 correct and 108 incorrect)

By far, there were two responses from the students that were the most common. These

were “pants that you wear in space” (or “to space” or “for space” or “from space”) and “pants

that astronauts wear” (or “pants for an astronaut”). Since these answers do not reference stars and

comprised over 50% of the unclear choices, the researcher made one observation of each of the

drawings paired with these answers. Simply, if the picture had a star or stars in it, the response

       

   

93  

was considered correct. Likewise, if there were no stars or other bodies instead such as the moon

or the earth, the answer was considered incorrect as there was not enough evidence to presume

conceptual understanding. The students’ pictures were also used in the same way when analyzing

the other responses provided. These definitions included “pants that are astro,” “pants that blend

into the sky at night,” “pants made in space and sent down to wear,” and “pants that take you to

the sky.” Each of these was accompanied by a picture that included one or more stars. Therefore,

they were determined to be correct.

There were other answers, however, that were identified as incorrect due to the absence

of any stars in the drawing. These were “pants with rockets on them,” “flying pants,” “pants with

no gravity,” “pants on (or “from” or “for”) the moon,” and “pants that have a telescope on them.”

While they were all relative to outer space, they did not definitively show a student’s

understanding of the concept of “stars.”

8. “millistars” (48 reviewed; 1 correct and 47 incorrect)

There was only one student answer that was considered correct after further examination.

This one said “real fast and real tiny stars – a millisecond fast and a millimeter in size.” The

reason this was accepted was because, though there was not a clear reference to a “thousand” (or

“thousandth”), there did seem to be an understanding that anything “milli” was at least much

smaller in relation to the larger unit.

All other responses had to be counted as incorrect. Most of these said “a star that is a

millimeter” (or “a millimeter tall” or ‘long”) or “small (or “little” or “tiny”) stars. While these

were similar to the first student response, none of these had both the word part and the concept of

its size; they were comprised of one or the other. This made it much harder to determine if the

       

   

94  

child accurately understood the concept. Again, the drawings provided did not offer any further

evidence. Most were simply pictures of stars and nothing else.

9. “symmetrishapes” (136 reviewed; 116 correct and 20 incorrect)

There was evidence that the children had discussed this concept in at least one class as

virtually all students had the same, or very nearly the same answer. Again, the determining factor

in whether they would be considered correct or incorrect came down to the picture with the

description. There were really only two responses out of the 70% that were considered unclear in

the first round coding. With very minor difference in wording, the most frequent answer was

“shapes with symmetry” (or “a line of symmetry”). Each child then provided a drawing with

shapes. Again, it was clear that this had been done with students since most had the identical

picture: a series of shapes with the “lines of symmetry” drawn in them. Whether the children

truly understood the concept or were simply re-creating an illustration that had been provided for

them cannot be fully determined. However, credit for correct responses was given to those that

had drawings with shapes that were truly symmetrical and had accurate lines of symmetry drawn

in them.

The remaining answers, even though they may have written “shapes that are

symmetrical” were considered incorrect if they provided a picture of asymmetrical shapes and/or

they did not draw the lines of symmetry on the shapes they provided. Again, while this prompt,

in particular, seemed suspect in terms of true student understanding due to the nature of the

drawings, reasonable consideration was made to determine this level of understanding.

10. “lunafood” (48 reviewed; 10 correct and 38 incorrect)

Similar to the previous prompt for “astropants,” there were many student responses that

were the same or very similar and had to do with “space.” For the same reason, these had to be

       

   

95  

further examined with particular attention to the drawings provided. As with the other example,

student responses that included the defining picture (a moon) were considered to be correct.

Those that did not include it were incorrect. Therefore, answers such as “food that is from space”

(or “for” space or “made in space”) that were paired with a clear depiction of the moon (or an

identifying label such as “moon” or “craters” or “crescent”) were marked as correct. Other

correct responses included “a cookie that glows in the dark” (with a drawing of a planet that was

labeled “moon with craters”), food that works in the night (with a drawing with a “moon” label),

and a picture of a moon eating food. “Food for astronauts” was also accepted as correct as this

represented a relevant attribute of the moon (Frederick, Frayer, and Klausmeier, 1969). Finally,

“food that glows” was correct. While it did not include a moon in its picture, it also represented a

relevant attribute of the moon (Frederick, Frayer, and Klausmeier, 1969). Therefore, it was

marked accordingly.

Following the same theory, if responses about food “from/in/on/for space” did not have a

picture of the moon or including the moon, it was considered incorrect. There were many more

that were deemed incorrect for this reason than were deemed correct. Other incorrect answers

were “food with stars,” “food that feeds the stars,” “space food,” “space ice cream,” “food

shaped like stars,” and a picture with a bubble that said “I’m eating a comet.” None of these had

a drawing with a moon in it to provide further insight that the student understood that “luna”

meant “moon.”

11. “amphibicat” (98 reviewed; 73 correct and 25 incorrect)

The majority of answers that resulted in a need for second cycle coding were variations

of “part cat and part amphibian.” It became important to further identify whether the children

were aware of what “amphibian” meant. Therefore, an examination of the drawings they

       

   

96  

completed was used to look specifically for any relevant attributes of the concept “amphibian.”

There were, indeed, several (44) that provided a picture depicting a cat with one of the following

features, or attributes, of an amphibian: scales, fins, webbed feet, and/or turtle shell. There were

other responses that included the cat in the water or with a thought bubble with a hybrid phrase,

such as “ribmeow.” These were all considered correct in the second cycle. Other acceptable

answers combined a mammal (cat) with an amphibian (lizard, frog, salamander, newt).

Those that remained incorrect were those that had drawings difficult to distinguish. While

the assessment is certainly not designed to measure artistic ability, in order to be considered

correct, there had to be evidence that a reasonable person would/could consider to be a depiction

of either an amphibian or a feature/attribute of one.

12. “solarexplorer” (114 reviewed; 39 correct and 75 incorrect)

There were a great number of student responses to this prompt that required further

review (59%). The vast majority of these answers were “someone who explores space” (or “solar

system” or “planets” or “galaxy”) with very minor variations to this wording. Since the word

“sun” was the key word, they could not be considered “clearly correct.” The drawing, once

again, would be critical in the second coding. Since a sun is relatively easy for a child to draw

and for a reviewer to recognize, this became the deciding factor. All of the pictures that

contained a sun (even in the presence of other “planets” or “stars”) were considered correct.

Other correct answers included “gives off energy,” “studies light and energy,” and “soars

through space.” Again, each of these had to be accompanied by a picture of the sun, which they

were.

Likewise, if a picture did not contain a sun in it, it was considered incorrect. Answers

included the most common phrases noted above, as well as “solar panels,” “explorer that has

       

   

97  

solar energy,” “an astronaut,” and “someone who studies planets.” However, since none of these

included the sun picture, they were marked as incorrect.

13. “terrahouse” (16 reviewed; 11 correct and 5 incorrect)

As a reminder, 70% of the students in this grade incorrectly identified the concept “terra,”

most confusing it with the word “terror.” Of the 16 students whose responses required further

review, several included “a covered house,” “a terra house,” and “a house made of (or “in”) a

terrain.” Those that were accepted as correct were again accompanied by a drawing depicting

dirt, rocks, and/or grass around a house.

The remaining five answers defined it as a “house in a tree” or a “tree house,” but they

did not include any ground or earth in their pictures. Therefore, these were identified as incorrect

responses.

Grade Five  1. “florabook” (5 reviewed; 2 correct and 3 incorrect)

Only five children provided answers that needed further review as most of them

answered this prompt correctly. One student wrote “a flora book,” but the picture with it had

flowers drawn all over the cover of the book. The second child whose answer was considered

correct had no writing. However, they drew a picture that clearly represented flowers.

The three remaining incorrect answers were “me reading a flora book,” “a book called

‘Flora the Queen’,” and “a florist reading a book.” While each of them provided a picture and

they may have understood the concept, none of the pictures identified any flower or plant life.

Therefore, it could not be concluded that the children did, in fact, understand this concept with

certainty.

       

   

98  

2. “biohouse” (72 reviewed; 40 correct and 32 incorrect)

The majority of responses that required a second coding review were due to the students’

uses of the prefix “bio” in other words. This facilitated the need to investigate further whether or

not they provided more details in their drawings. Students whose picture referred to the actions

of living things and then had clear pictures of living things in their pictures were considered as a

correct response. However, further scrutiny utilizing the indicators of conceptual understanding

was used. Responses that indicated that students were defining the term structurally were

definitively counted as correct (Frederick, Frayer, and Klausmeier, 1969). These included any

written responses (accompanied by the appropriate pictures) that detailed a house that was made

of living things. Such answers were “house made of biomes” (with plants, trees, and flowers),

“house made of trees,” “using flowers, trees, and grass to make a house,” and “house made from

nature, flowers.” There were still more students, again, with the appropriate pictures, who gave

evidence of defining the concept operationally (Frederick, Frayer, and Klausmeier, 1969). These

included, “house that holds and grows flowers,” “house that animals grow in,” “house that is

helpful to the environment” (flowers and trees), and “houses made for plants to grow in.” Three

students identified a non-example of the concept by defining it as a “biohazard,” and depicting

things inside of it not living (Frederick, Frayer, and Klausmeier, 1969). Finally, a few students

described the house by identifying relevant attributes as one that is “living” (Frederick, Frayer,

and Klausmeier, 1969). This included “a person house” and “movable house” which both had

pictures depicting the house as a face or a person.

The first cycle coding identified some student misunderstandings. There were more

examples of this in the second cycle coding. For example, children clearly thought of the word

“biography,” but not necessarily the concept “life” in their descriptions and pictures. This was

       

   

99  

done both explicitly such as “house that is a biography,” “house about a biography,” “house

where you get biographies,” and “house you do a biography in” as well as implicitly in answers

like “story of your house” and “all about me.” The other answers that were not considered as

correct after the second review were due to the fact that none of them included pictures that were

clarifying or provided any further information to show conceptual understanding. The written

responses included “a person saving the environment,” “a biologist’s house,” “a place to study

animals,” “nature’s house,” “a house in the family,” “a house in the forest,” “a house on earth,”

and “a dog house” (or “bird house”). However, none of the pictures included any living subjects

such as people, animals, or flora.

3. “omnicolors” (81 reviewed; 57 correct and 24 incorrect)

There were more students whose answers required further review than there were

students with clearly correct answers. Of the 81 students whose answers needed further review,

the majority of them (57) were later determined to be correct in second level coding based on the

drawing provided. Since the key for “omni” required the words “many/multiple” or “all” in the

written responses, many students did not have this exact wording. However, upon reviewing

their pictures, in context with the wording, it could be reasonably determined that there was

conceptual understanding of “omni.” For example, several students drew either a color wheel or

a rainbow, and all of them represented many colors on them. The written descriptions for these

included “wheel with random colors,” “sketch colors,” “tie dye rainbow,” and “rainbow with

different colors.” Other children used pencil shading and/or colored pencils and markers to create

“many” different colors and explained their pictures as “splots of paint,” “20 (or “30”) colors,”

“both hot and cold colors,” “all different kinds of colors,” “multicolors,” and “bunch of different

and crazy colors.” Other pictures represented multiple colors in both their pictures and their

       

   

100  

vocabulary. Examples were “different colors,” “use all the shades,” “primary colors mixed,” and

“color with different parts.” One student simply wrote “paint.” However, their picture was

several circles or “splotches” of paint. This was considered to be representative of “many,” and

thus marked correct.

The incorrect responses seemed to center on the same misconception as observed in the

first cycle coding. However, the students’ pictures did not provide any further clarification.

Therefore, they were considered incorrect after further review. Again, as in the first cycle coding,

it was recognized that students misunderstood “omni” to mean “eat.” Many answers referred to

this as “colors only omnivores eat,” “a bear is an omnivore and it is colorful,” “different colored

food,” “colors eating other colors,” “people that represent the colors they eat,” and “colors that

eat everything.” When it was still unclear after second cycle coding whether the child clearly

understood “omni” to mean “many,” it was marked as incorrect. One example was “an omnivore

with different colors on it.” While this student referred to “different colors,” the fact that they

also wrote about the “omnivore” made it unclear which was their definition of “omni.” Similarly,

when it was simply repeated, as in “omnivore colors,” there was not enough information to

consider the student’s intentions to be the correct answer. Other incorrect responses had to do

with the “omnivore” error, but more indirectly. These included “colors of food,” “a colorful

dinosaur,” “meat and plant colors,” “colors of nature,” and “colors based on plants.” There were

two responses that remained too vague, even after further review. One of these was “crayon with

omnicolors.” The picture with it did not include more than one color. The other answer was “the

colors are calming down.” While the student’s picture depicted three colors in it, this was not

determined to be enough evidence that the child’s focus was “many.” Finally, when a child

       

   

101  

referred to simply “two colors,” this was considered incorrect as well since, while it represents

more than one, this is not the same as “many” or “all.”

4. “trifeet” (1 reviewed; 1 correct and 0 incorrect)

Only one answer was reviewed for a clarification of the child’s depiction of the

morpheme “tri.” This student wrote “feet that are triangles.” Their picture did, in fact, represent a

three-sided foot. Therefore, this was marked as correct.

5. “electrodog” (54 reviewed; 33 correct and 21 incorrect)

Since 54 children did not describe this prompt with “electricity” or “electrical, their

pictures were reviewed and their wording was further reviewed. In order to determine if the

children understood that “electricity” was the correct concept, they had to have a relevant

attribute (Frederick, Frayer, and Klausmeier, 1969) of electricity in their written answer and/or

represented in their pictures. Therefore, if a child’s written answer was “robot(ic) dog,”

“electronic dog,” “dog that was struck by lightening,” “my dog is shocking me,” “dog who got

shocked,” or “dog who got electrocuted,” they had to be accompanied by a picture clearly

representing lightning bolts, an electric cord, an electric outlet, and/or an electric on/off switch in

order to be considered correct. This was the case with 33 of the answers that had to be reviewed.

All of the clearly incorrect answers, except for one, were worded as either “robot(ic)

dog,” “electronic dog,” or “remote control dog.” However, none of these had a clear indication

of electricity, including in the pictures they drew. Therefore, it could not be determined on this

evidence whether they understood the idea of “electricity.” The last answer reviewed, but

ultimately also marked as incorrect, simply re-stated the word by writing “electro dog,” and their

picture did not offer any further clarification.

6. “respihelmet” (14 reviewed; 5 correct and 9 incorrect)

       

   

102  

Again, the picture became extremely important in differentiating between which of the

reviewed answers would be considered correct or incorrect. Answers ultimately determined as

correct or incorrect may have had similar or even the same wording, so the picture and the

consideration of whether there were conceptual indicators were the determining factors. Answers

such as “helmet with respirator,” “respiratory helmet,” and “scuba helmet” were accepted as

correct if they included a picture of air bubbles or relevant factors (Frederick, Frayer, and

Klausmeier, 1969) that served as conceptual indicators such as a mouthpiece or someone

coughing. In terms of the incorrect answers, some of them had wording similar to those that were

considered to be correct; however, unlike those answers, they did not provide any further

information for clarification, including in their pictures. Other answers included wording that

may have indicated a relationship with breathing, thus requiring further review. However, none

of them contained any further clarification that would have been evidence of the child’s

conceptual understanding, so they were marked as incorrect. These included “space helmet,”

“gas mask,” and “sprayer helmet.”

7. “unistar” (24 reviewed; 18 correct and 6 incorrect)

Of the 24 answers reviewed, two of them provided wording and a picture that were

different than the others being considered again in second cycle coding. These were “only star in

the galaxy” and “a girl that became a star on her own.” Both of these responses were concepts

defined semantically (Frederick, Frayer, and Klausmeier, 1969), and therefore were accepted.

The remaining answers that were marked as correct were alike and involved many stars coming

together as one. However, if they did not include a picture that clearly depicted this, they were

not included. Such pictures included one large star with many stars inside of it or many stars

drawn with arrows all pointing to one main star. The wording included “the stars are coming

       

   

103  

together,” “a group of stars,” “a union of stars,” “stars joining together” and “all stars are

together.” There were other answers that were worded similarly. However, they were marked as

incorrect because they did not have the accompanying picture for clarification. Definitions such

as “group of stars” and “same stars together” did not count because they simply drew many

separate stars on the page. Other incorrect answers were “someone who’s not a star,” “universal

star” and “unicorn shaped like a star.”

8. “visiphone” (77 reviewed; 31 correct and 46 incorrect)

While some students did not use the key word(s) in first cycle coding to be considered

clearly correct, they did offer terms that were relevant attributes of “see” or “seeing” as well as

pictures, the most common of them being a set of eyes. Therefore, “a phone with vision,” “phone

with an eye,” and “an EYE phone” that included pictures of eyes were all marked as correct.

Student drawings were also helpful in determining that various wordings for “an iPhone for Face

Time” and “a phone that someone can visualize” were also accepted. These depicted people

“seeing” each other or “seeing” the phone. Another picture showed a person with a line of dots

moving from the person’s eyes to a phone with the description “me finding my phone.” A

conceptual indicator was also provided in the form of a non-example (Frederick, Frayer, and

Klausmeier, 1969). A few students provided answers referring to an “invisible phone.” Their

pictures depicted a phone present and then a space with a very light outline of a phone or simply

the word “ring.” These indicated understanding of the phone being seen and not being seen.

However, similarly worded answers were, once again, not accepted if this type of clarification

was not available in the student drawing. The majority of responses included wording such as “a

visible phone,” but drew a picture of a phone only. While they may have intended to show that

this was a phone that could be seen, without some other indication or evidence, this could not be

       

   

104  

assumed or accepted. Therefore, 40 such answers were marked as incorrect. The remaining

answers that were considered incorrect after review were all for the same reason. They provided

wording such as “phone that you can sense,” “invisible phone,” “big phone,” “this iPhone,” or

“phone that is a light,” but there was no further explanation in the picture that made it clear they

knew this meant “see.”

9. “chocolovore” (47 reviewed; 8 correct and 39 incorrect)

The only answers that were considered unclear and sent for second cycle coding were

variants of “someone who likes chocolate,” or “a person who loves chocolate.” The picture had

to provide the necessary clarification of whether the student was directly referring to “eating”

chocolate. Therefore, if the student provided a drawing that clearly depicted someone in the act

of “eating” chocolate, their answer was considered correct. If there was no such picture, or if the

student drew chocolate only, it was counted as incorrect. Using this method of second cycle

coding, there were only 8 correct answers, and 39 were incorrect.

10. “matricake” (19 reviewed; 11 correct and 8 incorrect)

There were only two types of answers that were considered correct after review. One was

when students provided the conceptual indicator of a semantic definition (Frederick, Frayer, and

Klausmeier, 1969). In these cases, students described, “mother cake” as “female cake leader of

the family,” “cake that rules other cakes and is a girl,” and “cake with a matriarch on it” that

included a drawing of a female on the cake. The other types of answer with conceptual indicators

were those that described “family” in place of “mother.” This is an example of a supraordinate

concept (Frederick, Frayer, and Klausmeier, 1969). Answers in this category included “”cake

that spells family,” “a family of cakes,” “a family cake,” “cake made by a family,” “a cake for a

family,” and “a candy made family on the top of a cake.” Each had a picture representing their

       

   

105  

description. There were also student responses that were changed to incorrect once reviewed.

These included “a girl cake,” “a cake that has control over others,” “woman on a cake,” “cake for

a woman,” “cake as a wife,” and “cakes that are related.” While each of these is similar to the

answers that were marked correct, they were missing important evidence. For example, it

contained the fact that it was a female or it was in control; not both a female and in control as

was the case with the correct answers.

11. “photofood” (37 reviewed; 18 correct and 19 incorrect)

Upon further review, there were student answers that were considered correct due to the

specificity of the paired drawing and/or the presence of a conceptual indicator. In this case, the

word “sun” or a picture of the sun and/or the sun’s rays, was considered a relevant attribute of

“light.” Correct answers included “food made by the sun,” “photo = sun,” “food for the sun,”

“food from the sun,” “sun cooked toast,” “food growing toward the sun,” and “food on the sun.”

Each of these written responses contained the word “sun” and provided a clear picture of the sun.

Other answers that did not specifically include the word sun were “a bush going through

photosynthesis,” “photosynthesis that makes food,” “a plant doing photosynthesis,” and a simple

restatement of the prompt, “photofood.” Each of these answers was accepted as correct in second

cycle coding because they were accompanied by a student picture of the sun that indicated

particular focus on the rays of the sun leading to the other part(s) of the picture that. This

emphasis was accomplished by the student drawing an arrow to the lines coming from the sun or

circling them, thus indicating the sun’s “light.” One more answer that was considered correct

was “holographic food.” This response was paired with a student rendering of a device

projecting lines to the subject. This was determined to be light, as with the drawings of the sun.

Finally, one other answer, “cheeseburger in a picture,” included the depiction of a camera with a

       

   

106  

flash. The student added an arrow pointing to the flash and wrote in the words “flash of the

camera.” This was considered to be an indication that the child understood the emphasis was on

the flash of light vs. the photograph itself. There were many other answers that could not be

determined to indicate student understanding of the concept, so they were marked as incorrect.

This included several (15) that were worded as “photo of food” or “picture of food.” What made

these responses different than those that were determined to be clearly incorrect was that they

included pictures that depicted a camera with a flash. While most emphasized the flash by

drawing lines from the camera, it could not be confidently determined that this was the focus of

the picture and a clear indication that the child understood that the core concept of “photo” was

“light.” Therefore, each of these responses was considered incorrect. Other answers that were

marked incorrect were for a similar reason. They did not include a picture that clarified whether

they clearly understood the concept. These were “food that goes through photosynthesis” (with a

picture that was not discernible), “photosynthesis” (with a picture of a flower), and two

responses that did not have a picture, which were “food made by an organism” and “food that

makes itself.”

12. “subdirt” (36 reviewed; 32 correct and 4 incorrect)

Several of the student answers provided clarifying pictures to assist in the determination

that their responses were correct. These included the written responses “dirt after the first dirt,”

“second layer of dirt,” “two layers of dirt,” “different name for subsoil,” “middle dirt,” dirt that is

deep in the ground,” “replacement dirt,” “dirt that is on the sub side,” and simple restatements

such as “subdirt” and “subsoil.” These were considered correct after review because each

included a picture that clearly indicated something “under” something else. This was done by

drawing arrows to the lower level or labeling layers, and showing the lower or lowest layer was

       

   

107  

the “sub” layer. One answer, “I dug so far I found subsoil,” was also marked correct. It was

determined that this was a conceptual indicator as it semantically described (Frederick, Frayer,

and Klausmeier, 1969) something that would be “deep” or “below” the surface. The answers that

were deemed incorrect may have had similar wording such as “two dirts,” “two kinds of dirt,”

and “different layers of dirt,” as well as a picture with multiple layers. However, they did not

label their pictures to show that their intention was to indicate the lower level. Therefore, these

answers could not be considered correct. One more incorrect answer was “sub that has dirt on it.”

However, after further review of the picture, the drawing indicated a submarine with dirt on it,

and there was no indication of the “sub” being “under” anything.

13. “duoflag” (16 reviewed; 14 correct and 2 incorrect)

The student pictures were once again extremely useful in determining whether responses

were correct in the second round coding. All of the student responses that were identified as

correct were deemed so due to a clear depiction of two items in the drawing. Written answers

included “a team of flags,” “the flags were a duo,” “more than one flag,” “looks like a lot of

flags,” “duo flags,” “a U.S. flag and a Massachusetts flag,” “a double flag,” “one flag with

another attached to it,” and “the American and Japanese flags.” Again, each one of these

included a picture of exactly two flags. The answers that were incorrect did not have clearly

distinguishable conceptual understanding in the student drawing. These responses were “flag

with more than one flag” and “flag with multiple things on it.” The former was paired with a

picture that had many lines and shapes on it. It was not clear that the intent was to represent only

two. The latter was combined with the representation of a flag with the words “we win,” and “we

lose,” printed on it by the child. While this represents two phrases, it was considered to require

       

   

108  

too much inference to definitively conclude that the child meant to communicate the concept

“two.” Therefore, these responses were marked as incorrect.

Additional Data   In both grade four and grade five, it was observed that students utilized a strategy on their

worksheets that was not solicited. This was breaking the prompt word down into parts by the

root and the compound word with which it was combined (“separation” strategy). For example,

if the prompt was “trifeet,” students underlined, circled, or drew a separating line between “tri”

and “feet.” This was seen more frequently in grade four. However, the grade five students who

utilized this strategy did it far more successfully. Figure 4.19 illustrates these percentages for

each grade that made use of this strategy correctly and incorrectly.

Figure 4.19: Percentage of correct and incorrect responses when students used the “separation”

strategy.

Correct responses Incorrect responses

Grade 4 46% 54%

Grade 5 71% 29%

Grade Level Focus Groups   Focus groups were conducted for each grade level and each content area. The special

educators for each grade level were invited to join the teacher groups for the focus group

sessions. One of the grade four classroom teachers (ELA) and one of the grade five classroom

teachers (ELA) declined to participate. All other classrooms teachers were present (eight) and

special educators were also present. Therefore, the following focus groups were held: grade four

ELA, grade four science, grade five ELA, and grade five science.

       

   

109  

The researcher offered the administration and the teachers the choice of meeting time and

whether they preferred after school or during the school day. Everyone agreed that the best and

most convenient time for participants was during the teacher preparatory periods during the

regular school day. They were further given the option of focus group sites. The intention was

for the focus groups to be in an environment that was familiar and comfortable. It was also

important that there be no ambient or background noise so that the recordings could be

completed successfully. Each focus group took place in a classroom or teacher meeting room,

and participants, including the researcher, sat around a large table so that everyone could see and

hear one another easily. Teachers were reminded that they would be recorded, and the recording

device, an iPad with the Rev.com App, was placed in the center of the table.

Each of the focus group sessions was completed within 30 minutes. These sessions were

conducted following the focus group model by Krueger and Casey (2009). The same questions

were presented to each group and they were open-ended, to eliminate interviewer bias. They

began with opening and introductory questions, progressed to transition and key questions, and

concluded with ending and follow-up questions (Krueger and Casey, 2009). All focus group

questions are found in Appendix J.

Each participant was told that members could be interviewed separately and at a later

time if they wanted to share further information in the event they wanted to speak privately or

simply remembered more information to share. None of the participants exercised this option.

Each person in the Focus Group contributed to the discussion. Focus Group flaws were reduced

by utilizing the strategies of Krueger and Casey (2009). The interviewer was able to keep all

participants on task, have everyone contribute at least once, and avoid any one participant from

dominating the conversation. In addition, the interviewer provided a brief oral summary after key

       

   

110  

points were made, and before concluding each Focus Group, also allowed a moment for any

teacher to add a comment, have a question answered, or modify a prior response.

After the transcripts were professionally transcribed by Rev.com, electronic versions

were provided to every Focus Group participant. They were prompted to correct any errors in

meaning or intent. They were further encouraged to clarify any inconsistencies or provide

clarification if needed. There were no corrective responses from members of the Focus Groups,

so coding began. This fact-checking step was important to maintain the integrity of the study and

the accuracy of the participant responses.

First cycle coding for focus groups. Analysis was completed utilizing the Classic

Approach (Krueger and Casey, 2009). Every transcript was read thoroughly once. They were

then read a second time, this time color-coding the major areas represented in the question

design: student growth and achievement (green), fidelity to the program and strategies used

(yellow), and student difficulties and challenges (red).

Transcripts were read and details extracted in this order: grade four ELA, grade five

ELA, grade four science, and grade five science. This order was used purposefully so that the

researcher might be “alert to changes that may be occurring from one type of audience group to

another,” which, in this case, was being aware of any differences and/or similarities between the

fourth grade and fifth grade of each subject (Krueger and Casey, 2009).

The first area of analysis from each group consisted of any quotes that were indicators of

student growth or achievement. These were highlighted in green in the transcripts. The second

set of quotes extracted from the transcribed text were those that described any strategies or

alluded to fidelity or lack of fidelity to the program. These lines were highlighted in yellow.

Finally, the teacher quotes that indicated a difficulty or challenge with the program and

       

   

111  

morphological study of vocabulary were highlighted in red. The result of all of these quotes is

represented in the following charts, followed by a description of the second cycle coding for

thematic connections.

Grade 4 ELA Focus Group QUOTES - Cycle one coding Growth/Achievement

Green Fidelity/Strategies

Yellow Difficulties/Challenges

Red • I like it only because I

took Latin for a few years…I get the correlation

• I think there’s more (vocabulary)

• They remember the root…they can remember more words and at least be able to know how to break it down

• Yes (students are noticing word parts without prompting)

• A lot of the time (they know what the word part means)

• This group definitely has higher fluency and comprehension

• I think they’re more aware that they can break down prefixes and suffixes in the Greek words to help them

• Not now, but at the beginning of the year (had to prompt students to use strategies)

• I have a couple in my class when you say Latin…they get very excited

• They love using (flashcards)

• I’ve been going through the workbook page-by-page

• That’s the only resource I have (workbook)

• Few fold-ables • We don’t give them

quizzes and tests on just those pages

• It’s not every week… not necessarily what we’ve been doing in the workbook

• They have flashcards

• No collaboration with science teachers on common vocabulary

• Overwhelming at first • Lack of resources • Hard to squeeze it in with

other things • Very hard to fit in with

everything else • With time, it’s just not

feasible • I don’t find there’s very

good mastery • Just more exposure right

now…I feel like it’s a drive by

• I personally think they’re getting less (mastery)

• We don’t spend enough time on things

• I wouldn’t say 100% across the board (that students are using strategies on their own)

• If we had an actual program or different activities we could do with them to make them own them more

       

   

112  

• Helps with syllabication (they chunk it based on the word part)

Grade 5 ELA Focus Group QUOTES - Cycle one coding Growth/Achievement

Green Fidelity/Strategies

Yellow Difficulties/Challenges

Red • Once you start using it,

the conversations that happens with the kids, they make a connection that it all wraps together

• Integer…they brought that into math

• (Students) were excited to inform the teacher that they knew all about the word before she even spoke about it

• (Students) are happy that they know that (root meaning)

• They were making the math connections during ELA class

• (Students) were noticing it before we were instructing it

• Starting early January…(students) started (noticing connections)

• Kids would make that connection without doing a structured Greek and Latin…lesson

• They also made that connection, ‘Oh, we learned that in science’

• Yeah (they understand the concept)

• It’s extended beyond the academics

• Introduced maybe three years ago (Prestwick House)

• After the fact, we (teacher pairs) have some conversation (about shared vocabulary)

• The units were set up originally like as to who was going to teach what part…(we) met…and we kind of broke the unit up…more of an informal conversation

• Conversation in the beginning

• Showing them the root and having them activate their thinking …do they know any words with those pieces

• Put the notes on the board and (they) copy it into their notebooks so they have a chapter on each Latin root section

• I do charts…from the beginning of the year (with) all of the Latin roots

• I also take notes (for) the kids who might

• There’s no official common planning

• I was intimidated by the time and also for the children who are struggling with reading in general…I found it was too much

• Initially it felt like it was a separate program

• We instructed (students) more in the beginning of the year

• Time wise it’s still very difficult to get it in

• When (students are) generating a lot of words that actually don’t have a Greek and Latin root

• I’ve found that to be a stumbling block with just generating words that they think have (roots)

• The kids who are lower really struggle with making the relation to it

• I haven’t tried to do a lot with the Greek and Latin with (lower achieving students) because they’ve got so many other priorities

• Just find the whole time thing really difficult because there’s so much (to cover)

       

   

113  

• A student who has historically struggled, when we did ‘urb,’ for ‘in the city,’ we have a teacher in the building who has that as part of her last name and he referred to her as ‘Mrs. In-The-City’

• My kids made that connection

• I’m much more comfortable

• It comes more natural to the kids now. It’s a way of life

• They are starting to make connections and point them out to me with the roots

• Even if they come across something in the reading and I’m not noticing it, they are…they are making connections

• We know the units (the science teachers) teach…or the concepts that they cover

• The kids will point it out • It feels (they are) more

actively acquiring them the way we’ve done it this year

• There’s more of a connection to other words because of the similarities

• Yeah (they respond when relating roots to other words)

• I think some of them do make the connection

• They’ll point out…the Latin word and they want

miss (it) or are absent • If they have trouble

copying…I’ll give them my copy

• They’ll work collaboratively…and then sometimes we give them independent work on the workbook pages

• We do more finding words

• Writing sentences • Word spokes • I do a chart where

they find (words) that have (roots) in the beginning of the word or in the middle or the end

• They do sentences in a picture or on circles (Frayer Model)

• Mine tend to use Greek and Latin dictionary…they’re familiar with how to find that

• They write sentences • Kids pull words out of

their pleasure reading • Routine of Greek and

Latin roots • We do a drum roll

• Given the high stakes test, it would’ve been nice to have (covered roots earlier in the year)

       

   

114  

to show you that they see it and that they know what it means

• Definitely (they’re showing even when not asked)

• They’re excited that they’ve made a connection…and they’ve noticed it so they’ll bring it to our attention

• They’re more familiar with it (due to routines) and they notice it more

• It’s like ‘Oh, it’s cool so I’m going to point this out’

• They like to show off • Honestly, I enjoy

them…when they point it out it makes me happy

• It’s that recognition…’I know (the teacher) likes it and I found one…I’m happy’

• This is making me look good (students feel)

• Kids do a drum roll and they like that

• (It’s) like a self reward • The boy with ‘Mrs. In-

the-city,’ he definitely has a lot of challenges and that was excitement for him, he was motivated, definitely

• When they try to articulate it they say a lot more than what they show (in tests)

• The more comfort I (teacher) feel with it, the more I see (students’) comfort as well

       

   

115  

• I see (students) making connections and growing above and going beyond that

Grade 4 Science Focus Group QUOTES - Cycle one coding Growth/Achievement

Green Fidelity/Strategies

Yellow Difficulties/Challenges

Red • I was overhearing (a

student)…they wanted to know how many sides an octagon had and the other student said, ‘well, octo, octopus.’ They were trying to help that child make a connection.

• Probably more so in math they make those connections

• I don’t focus on them (Greek and Latin)

• Sometimes (roots) come up in measurement a little bit, but not to really focus on…we talked a little bit, but very brief

• Word splash, flashcards, fill-in-the blank…sentences with the word bank on the top

• A matching game where they have to go around and find their match

• Part of the test is vocabulary (embedded in the regular unit test)

• We might have to start working on the aligning…we are going to have to do something different soon…what that is, I really don’t know

• We’ve also talked about reordering the units and changing how we teach things…we’ve talked about integrating

• No (teacher collaboration between ELA and science)

• We pretty much do our own (vocabulary)

• It’s a million years old (science materials)

• No (not aligned to new standards)

• No (don’t know the word parts children learn in ELA)

• Now we’re more departmentalized, it’s more difficult to incorporate ELA and social studies

• We do (have common planning) with our partner…but you don’t really go across (to the other teachers or other grade)

• No (we don’t point out roots)

• Right (the textbook doesn’t make connections) the only prefix that we’re doing (is) vertebrate and invertebrate, but that’s it

• No, they’re not (making connections to words done in English class)

       

   

116  

some of the math into the science

• If I’m teaching…I will mention the fact that they come from the Greek language

Grade 5 Science Focus Group QUOTES - Cycle one coding Growth/Achievement

Green Fidelity/Strategies

Yellow Difficulties/Challenges

Red • Sometimes (students say a

word part looks familiar) • (Students) will usually say

they saw it in ELA class or they are familiar with it in some other reading

• When we came up with hydrotropism…the kids said ‘hydro, that means water,’ so they knew that

• Yes (it may be helpful in understanding concepts)

• Tropism(s) are the biggest ones…they do ‘geo’… ‘photosynthesis,’ and words like that

• Students usually tell me (which word parts they know)

• It’s my higher kids that I see sometimes pulling (a)part words or making that relationship

• If they can make that connection and know the parts, the meanings of the parts of the words, then they do understand it better

• One of the students made the connection (when doing fractions) that ‘uni’ means one…having a one

• If (Greek or Latin word parts) come up in a word, like in the word ‘photosynthesis,’ I’ll point it out: ‘It has photo, which means light’

• I don’t ever specifically say it’s Greek or Latin (but) I do the same thing (break the word down)

• I might (identify Greek or Latin) if I know which language it comes from

• I’ll underline it so I’ll specifically bring it to their attention

• I don’t ever reference whether (a word) is Greek or Latin. We just pull apart the word

• I do it sometimes (use a word attack strategy)

• My partner has a list in her room (of word parts in ELA), so if I’m doing something

• I don’t ever specifically say (a word) is Greek or Latin

• No, not too much (collaboration with the ELA teacher)

• No (there are not a lot of words that repeat in ELA and science)

• I don’t think we all probably have the same (vocabulary)

• Not always the same time (do we do the same units)

• If I do (notice any difference in a student’s level of comprehending the concept) it’s a select few

• There’s not many (other differences noticed)

• Students find (my end of year Greek and Latin activity) difficult and say that they find it challenging, most of them

• No one’s really ever mentioned anything to me (about understanding a word because of knowing the word part)

• I don’t think that the majority of kids have that

       

   

117  

in the numerator…they made that connection themselves

• I had someone say that, too, when that came up (uni)

• If they know it, they are kind of excited about it

• I find mine’s working fine. I usually do pretty well on vocabulary

in science I’ll go and put it on her list so the students will see it in both places

• I’ll put it on her board so that she sees or the students will see the same word in my class and their class

• We use a lot of supplements, things that we find

• If it’s in the frameworks (the vocabulary) will be exactly the same (in ELA and science)

• I know they do that one in ELA (‘uni’) from last year

• I do an activity that has Greek and Latin roots…at the end of the year

• They have activities so all the ‘hydro’ words or all the math words like ‘uni’ or ‘quad’…they have some activities where they need to use dictionaries and the students find that difficult and challenging

really good understanding of it. They can tell me the definition but to use it and apply it, they lack. Especially when you move on and talk about it…it’s not there anymore

Second cycle coding for focus groups. A second cycle coding was needed to identify

categories and/or themes. These could not be pre-determined without knowing the participants’

responses. There are four factors recommended by Krueger and Casey (2009) for use when

assigning weight to themes. The first is frequency. This takes into account the number of times a

       

   

118  

comment or point is repeated while still acknowledging the importance of the content to the

topic. The second factor to consider is specificity. It is suggested that the more detail provided

about a point, the more weight it should have in the development of themes. Emotion is the third

factor. This is when a participant has exhibited “emotion, enthusiasm, passion, or intensity” in

their answers (Krueger and Casey, 2009). Finally, extensiveness was considered. This goes

beyond frequency, or merely the number of times something is repeated. Extensiveness includes

the number of different people who say the same thing as well.

Growth/Achievement  

Each of the categories identified in first cycle coding was examined independently and by

grade level. The growth and achievement comments yielded five themes for both grade four and

grade five. These were student motivation and engagement, connections to other subjects,

conceptual understandings, student skill development, and teacher engagement.

The first to be described in more detail is the student motivation and engagement

theme. This was identified as a theme due to the frequency of times it was noted, the

extensiveness in that it was repeated by many people, and the emotion with which several

teachers communicated the information (Krueger and Casey, 2009).

In grade four, teachers noted this using the Greek and Latin roots when they reported that

students, “love using (flashcards),” and another noted that she had a couple students in her class

that, when she says “Latin…they get very excited.” Another said that students “were excited to

inform the teacher that they knew all about the word before she even spoke about it,” and

“students are happy that they know that” (root meaning).

       

   

119  

In grade five, the student motivation and engagement comments included that “they’re

excited that they’ve made a connection,” “they like to show off,” “if they know it, they are kind

of excited about it,” and it is “like a self reward when they find a root in a larger word.” Teachers

said the students “point out…the Latin word and they want to show you that they see it and that

they know what it means.” “It’s that recognition… ‘I know (the teacher) likes it and I found

one…I’m happy’.” “Kids do a drumroll, and they like that.” They add that the students are

“excited that they’ve made a connection…and they’ve noticed it so they’ll bring it to our

attention.” “It’s like ‘Oh, it’s cool so I’m going to point this out’… ‘This is making me look

good’ (students feel).” When a student made a connection to the teacher, she said “that was

excitement for him, he was motivated, definitely.”

The second theme was connections to other subjects. The specificity that teachers used

to describe the connections that students made in other areas and the frequent number of times it

was mentioned are the elements that contributed to this becoming a theme.

The grade four teachers observed that the students make connections to vocabulary in

other subjects, but “probably more so in math.” One teacher noted that she overheard two

students talking. One child “wanted to know how many sides an octagon had and the other

student said, ‘well, octo; octopus.’ They were trying to help that child make a connection.”

In the fifth grade, the frequency of the number of examples of student connections

increased. There were several illustrations from math. They said, “they were making the math

connections during ELA class,” “integer…they brought that into math,” and “one of the students

made the connection (when doing fractions) that ‘uni’ means one…having a one in the

numerator…they made that connection themselves.” The same is seen in science and outside

reading. Students say “Oh, we learned that in science” or “they saw it in ELA class or they are

       

   

120  

familiar with it in some other reading.” “If they come across something in the reading and I’m

not noticing it, they are…they are making connections.” They also note “It’s extended beyond

the academics.” “A student who has historically struggled, when we did ‘urb,’ for ‘in the city,’

we have a teacher in the building who has that as part of her last name and he referred to her as

‘Mrs. In-The-City’.” The teachers observe thdat “there’s more of a connection to other words

because of the similarities,” and that “sometimes (students say a word part looks familiar).”

Sometimes, “kids would make that connection without doing a structured Greek and

Latin…lesson.” “They are starting to make connections and point them out to me with the

root…once you start using it, the conversations that happens with the kids, they make a

connection that it all wraps together.”

Conceptual understandings is the third theme. This topic was identified in each session,

so the extensiveness is what qualified this as a central theme.

Teachers in grade four noted specific indicators of conceptual understanding. This

included stating that, “this group (of students) definitely has higher fluency and comprehension,”

“they’re more aware,” and they require less prompting to use the strategies. They also describe

children working to teach other children to use the strategies to understand words.

The teachers in grade five notice that the students seem to understand new concepts better

when they start with an understanding of part of the word. The say the process “comes more

natural to the kids now. It’s a way of life.” Another teacher said, “If they can make that

connection and know the parts, the meanings of the parts of the words, then they do understand it

better.” One teacher felt the student tests did not show the level of their full understanding. She

said, “When they try to articulate it they say a lot more than what they show (in tests).” Students

are “growing above and going beyond.”

       

   

121  

Another theme identified was student skill development. Teacher comments were

detailed, and this specificity, along with extensiveness of responses, is why this fourth theme was

chosen.

Students in grade four were described as being able to “remember the root…they can

remember more words and at least be able to know how to break it down.” They say that

students are “more aware that they can break down prefixes and suffixes in the Greek words to

help them.” A special educator noted, “it helps with syllabication” as well.

Several fifth grade teachers also spoke about “breaking down words” and the fact that

students are doing this more on their own now that they have practiced these skills with them.

This would happen even “without doing a structured Greek and Latin…lesson.” They felt that

the students were “more actively acquiring them the way we’ve done it this year,” “they’re more

familiar with it (due to routines) and they notice it (Greek and Latin word parts) more,” and they

sometimes see them “pulling (a)part words or making that relationship” on their own.

The fifth and last theme of student growth and achievement was the teacher

engagement. The extensiveness of these comments as well as the emotion that was attached to

them are what made this theme central to this category.

A fourth grade teacher said “I like it…I get the correlation.” Grade five teachers had

more specific comments. One said, “Honestly, I enjoy them…when they point it out it makes me

happy.” Another found that the students know that the teacher enjoys it. She said students like

the “recognition… ‘I know (the teacher) likes it and I found one…I’m happy’.” Two others

commented that, “The more comfort I feel with it, the more I see (students’) comfort as well,”

and “I find mine’s working fine. I usually do pretty well on vocabulary.”

       

   

122  

Difficulties/Challenges  

There were also five themes identified under the difficulties and challenges comments

made in the focus groups. These themes were lower achieving students, program coordination,

time/resources, level of mastery, and teacher mindsets/capacity.

The first theme, lower achieving students, was chosen due to the specificity of

comments in relation to student skill deficiencies.

A teacher in fourth grade said, “for the children who are struggling with reading in

general…I found it was too much.”

Fifth grade teachers said “the kids who are lower really struggle with making the relation

to it, and one said, “I haven’t tried to do a lot with the Greek and Latin with (lower achieving

students) because they’ve got so many other priorities.”

The next theme was program coordination. This was mentioned many times in each

group, so frequency and extensiveness were factors in choosing it.

In grade four, the ELA teachers reported that there was “no collaboration with science

teachers on common vocabulary.” The science teachers said there is no alignment to the Next

Generation Science Standards and no understanding of the word parts that the children learn in

their ELA classes. They stated, “We pretty much do our own (vocabulary).”

The fifth grade teachers noted that, “there’s no official common planning.” One felt as

though she instructed students “more in the beginning of the year.” One science teacher said, “I

don’t ever specifically say (a word) is Greek or Latin.” They agreed that there was “not too much

(collaboration with the ELA teacher).” They felt that they “don’t think (they) all probably have

the same (vocabulary),” and units are “not always (taught at) the same time.”

       

   

123  

Time/resources is the third theme. In addition to the number of various points made

about this, there was an element of emotion as teachers described this topic. Therefore,

frequency, extensiveness, and emotion were all reasons this was made a major theme.

The teachers in the fourth grade focus groups felt strongly about both time and resources

in their classes for this type of vocabulary work. In terms of time, they repeated this frequently,

saying, it’s “hard to squeeze it in with other things,” it’s “very hard to fit in with everything

else,” “with time, it’s just not feasible,” and “we don’t spend enough time on things.” They said

that it is “just more exposure right now…I feel like it’s a drive by,” and “now we’re more

departmentalized, it’s more difficult to incorporate ELA and social studies.” They said that they

had a “lack of resources,” and if we had an actual program or different activities we could do

with them to make them own them more,” that it would be more successful. The science teachers

said that their materials are “a million years old,” and the textbook doesn’t make the Greek and

Latin connections for them.

In grade five, the teachers made no mention of resources. However, they said that, “Time

wise, it’s still very difficult to get it in. One teacher said that she “just find(s) the whole time

thing really difficult because there’s so much (to cover),” and “given the high stakes test, it

would’ve been nice to have (covered roots earlier in the year).”

The fourth theme was level of mastery. Here, teachers made notes of depths of

understanding. There was specificity to these comments, which is why it was chosen even

though it may have not been stated as frequently as other topics.

In grade four, teachers said “I don’t find there’s very good mastery,” “I personally think

they’re getting less (mastery),” and “I wouldn’t say 100% across the board (that students are

       

   

124  

using strategies on their own).” A science teacher said that she did not notice that students were

making connections to words done in English class.

One fifth grade teacher found that she had some students who had difficulty

understanding why some word parts do not have the same meaning in all words. In science, a

teacher said, “If I do (notice any difference in a student’s level of comprehending the concept)

it’s a select few.” Another said, “I don’t think that the majority of kids have that really good

understanding of it. They can tell me the definition but to use it and apply it, they lack.

Especially when you move on and talk about it…it’s not there anymore.”

Finally, the fifth theme was teacher mindsets/capacity. This category incorporates how

teachers expressed feeling bout the Greek and Latin program as well as how it may have affected

them. The emotion expressed, combined with the frequency of comments, made this appropriate

to choose as a theme.

In addition to the challenge they expressed with time, fourth grade teachers said that the

approach itself was “overwhelming at first.” One said that she doesn’t point out the roots to

students at all, and another teacher commented that the science textbook that they use does not

make the language connections for them.

Time was also a concern for the fourth grade teachers. One said, “I was intimidated by

the time,” and another added, “Initially, it felt like it was a separate program.”

Fidelity/strategies  

The final code in first cycle coding was fidelity/strategies. Both grade four and grade five

noted many specific strategies that they used in teaching a morphological approach to

vocabulary.

       

   

125  

The grade four strategies for learning vocabulary in ELA included “going through the

workbook page by page,” using “a few foldables,” and “flashcards.” They said that they don’t

give quizzes every week, and when they do, they are not “on just those pages,” and “not

necessarily what we’ve been doing in the workbook.” In science, they may talk about the roots,

but it is “a little bit, but very brief.” One teacher said she doesn’t focus on them (Greek and

Latin),” and another said, “Sometimes (roots) come up in measurement a little bit, but not to

really focus on.” They also use “word splashes, flashcards, fill-in-the blank…(and) sentences

with the word bank on the top.” One teacher uses a “matching game where they have to go

around and find their match.” Vocabulary is embedded into the regular unit test. While they have

“talked about reordering the units and changing how (they) teach things…(including) integrating

some of the math into the science,” this has not yet been done. They acknowledge that they

“might have to start working on the aligning.” “We are going to have to do something different

soon…what that is, I really don’t know.”

The strategies that were used for teaching vocabulary in fifth grade ELA included

“Showing (students) the root and having them activate their thinking,” putting “the notes on the

board and (they) copy it into their notebooks so they have a chapter on each Latin root section,”

keeping “charts…from the beginning of the year (with) all of the Latin roots,” “finding words

(with the same word parts),” “word spokes,” “writing sentences,” pulling “words out of their

pleasure reading,” and using the “Greek and Latin dictionary…they’re familiar with how to find

that.” They also “take notes (for) the kids who might miss (it) or are absent,” or provide a copy

“if they have trouble copying.” Students will also “work collaboratively…and then sometimes

(they complete) independent work on the workbook pages.” A teacher keeps a chart where they

“find (words) that have (roots) in the beginning of the word or in the middle or the end,” and

       

   

126  

another has students “do sentences in a picture or on circles (Frayer Model).” Overall, students

follow a routine for Greek and Latin study. They said that, in terms of coordination, “the units

were set up originally like as to who was going to teach what part…(we) met…and we kind of

broke the unit up…more of an informal conversation.” For science, they said that, in terms of

strategies, they said they “use a lot of supplements, things that (they) find.”

Two of the science teachers in grade five said that they “don’t ever reference whether (a

word) is Greek or Latin. (They) just pull apart the word.” The others said that they did.

Specifically, one noted, “If (Greek or Latin word parts) come up in a word, like in the word

‘photosynthesis,’ I’ll point it out: ‘It has photo, which means light’.” Another says she might

identify the origin if she “knows(s) which language it comes from.” One teacher said, “I

underline it so I’ll specifically bring it to their attention.” The other science teachers agreed that

they know the list of ELA words with Greek and Latin word parts. Some make use of this to

coordinate with their ELA colleague. “My partner has a list in her room (of word parts in ELA),

so if I’m doing something in science I’ll go and put it on her list so the students will see it in both

places…I’ll put it on her board so that she sees or the students will see the same word in my class

and their class.” They listed a few examples of word parts that they knew they had in common

such as “uni” and “hydro.”

Individual Teacher Interviews  

The researcher utilized the Responsive Interview Model for the individual teacher

interviews (Rubin and Rubin, 2012) to extract the most authentic information from participants.

Questions were open-ended to allow for teacher elaboration and anecdotal information and

reduce interviewer interference in their responses. The responsive interviewing approach allowed

       

   

127  

the interviewer to add clarifying questions when and where necessary if there was an

unanticipated “wisp of insight,” or to “track down a new theme” (Rubin and Rubin, 2012).

Individual interviews were conducted with the Curriculum Leaders. These teacher-

leaders serve as the content area peer leaders for both grades four and five. The English

Language Arts (ELA) Curriculum Leader and the Science Curriculum Leader both happen to be

grade five teachers, but they are responsible for overseeing the curriculum and instruction in

their respective content areas for both grades. These are positions for which they apply each year

and receive a small stipend. Both of these Curriculum Leaders are in their third year of serving in

this capacity within the district. They receive specialized training in curriculum design as well as

content area pedagogy. They also meet regularly with the other Curriculum Leaders in the

district, PK-12, representing all content areas.

The Curriculum Leaders were interviewed separately. They also chose to have their

interviews conducted during the regular school day, and each was interviewed in their own

classroom. Both interviews were completed within 30 minutes. They were asked Tour, Main,

and Follow Up questions (Rubin and Rubin, 2012). Probes were also used when appropriate,

based on the participants’ responses. Each was reminded that they would be recorded, and an

iPad was placed in the center of the table between the researcher and the Curriculum Leader.

Both teachers presented comfortably and had the opportunity to answer every question fully and

to their satisfaction.

Electronic versions of the professionally transcribed transcripts (by Rev.com) were sent

to each Curriculum Leader. Each was prompted to review them for accuracy. Neither of the

Curriculum Leaders made any corrections. The Focus Group and interview transcripts will be

       

   

128  

maintained for the mandated period of time and then will be destroyed to protect confidentiality.

Participants were informed that this would be the protocol.

First cycle coding for interviews. Again, the open-ended responses of the Curriculum

Leaders were analyzed through fist cycle and second cycle coding. The first cycle consisted of

the same three categories that were used in the Focus Group analysis. These included student

growth/achievement (marked on the transcripts in green), fidelity/strategies (coded in yellow),

and difficulties/challenges (indicated by red).

The result of all of these quotes is represented in the following charts, followed by a

description of the second cycle coding for thematic connections.

ELA Curriculum Leader - Cycle one coding Growth/Achievement

Green Fidelity/Strategies

Yellow Difficulties/Challenges

Red • The kids will say, ‘Stop!

That’s one of the Greek and Latin words that we had’

• I always do something like, ‘Oh my goodness, I have no idea what…flourish means,’ and they go ‘Flor, flor! That means to thrive, grow well’

• They make the connections; I see them connecting to science

• ‘Vor, to eat,’ they went through ‘omnivore, herbivore,’ they made that connection to (the science teacher’s) class

• They’ll say to me, ‘Oh, we learned that in science’

• When we talked about ‘lat’ and ‘long,’ they were

• I would present the vocabulary; I’d have them look at the board to see what they thought they saw in common until we came to the root word, and then I would tell them the…Greek and Latin words that they came from and we’d talk about the definition

• Spend time trying to come up with words that we knew

• We look them up; do some application of it

• They had to find several examples of other words that they found the root in

• They wrote sentences

• The low, low kids have trouble making connections…to trying to find roots on their own and be able to apply them

• Overall the students are so low that they’re not making that connection form the year before

• I’m like, ‘Didn’t you do this last year?’ ‘No,’ some of them said…word parts

       

   

129  

like ‘Oh, like the latitude and the longitude that we did?” They kind of make the connections… maybe more so in science, but I think all the way across I see the connections

• They remember what the words mean

• It’s filled their vocabulary with words that were provided as well, and they’re taking when words are out of context they’re able to put the words together by using the roots

• We were talking about the branches of government and they go, ‘Preamble, pre…that means before. That was the introduction to the Constitution!’ They’re making those connections

• They do refer back to science, like ‘bi’ for ‘biome’

• I see home connections like ‘dorm, sleep.’ (Students say) ‘Oh, my sister lives in a dormitory’ or ‘My brother is in a fraternity.’ They’re making real life connections as well

• There’s all these ‘ah ha’ moments where the kids will say, ‘That’s Greek or Latin’ or ‘That’s a Latin root’

• Kids will come in, they go on their own, they love to find words that have roots in for the

with them; they wrote stories with them

• We discussed them indirectly oftentimes with reading selections

• I…model the lesson • I made myself a few

notes (of science terms)

• I also find the affixes very helpful as well, because we do that weekly

• I see it right there (connect with social studies)

• (There’s) a set time during the day that we’ll say every single day that we talk about (vocabulary)

• Ten (words per week) • (Student performance)

all depends on the fidelity of the program

       

   

130  

week…sometimes they’ll say, ‘You know what I saw on TV…?’

• They’re excited about it • The teacher on my team

(has ‘urban’ in her name). My kids are like ‘Mrs. In-the-city…oh, she lives in the city’…it’s one of their special ed kids

• They learn so much from it. I can’t stress enough how much they continue to use those words of the week; they’re now embedded in their vocabulary

• Absolutely. Yep, absolutely (it helps them understand concepts)

• They do, all the time (hypothesize the meanings of unknown words)

• When there’s no context clues at all…they’re uncovering new meanings…it’s not always accurate, but they attack it.

• They try to outdo each other…it creates competition

• They have to find some (words) on their own, and some go way above and beyond, like they like it; it’s a challenge for them

• Very engaging…this year the kids are somewhat low overall, and they’ve just done amazing with them (roots)

• They’ll say ‘this is fun. That’s a good way to do

       

   

131  

this. I never thought of this before.

• This group was really eager to make those connections and they did it on their own

• In general; all of (the students) make those connections

• As a result…they have built and they’ve acquired new words, but I think they’ve built an admiration for learning new words…they look forward to it

• They want to build, they want more words, they want to be able to use them and show them off

• They’re just so excited about it

• Absolutely (it gives a sense of pride). One of our themes was ‘How does something inspire you?’ Right away they kept referring to ‘breathing it in’

• I think it’s a really great program

Science Curriculum Leader - Cycle one coding Growth/Achievement

Green Fidelity/Strategies

Yellow Difficulties/Challenges

Red • We’re using the same unit

(the science teachers) • Correct (most likely

choosing the same vocabulary words

• I have a list of them (ELA words)

• Yes (all science teachers

• PowerPoint; I’ll put up the word. We might talk about it ahead of time

• If I think they’re already familiar with the word I might ask them, ‘Do you know

• We’re not using NGSS yet in our instruction

• I put it (vocabulary) as part of my unit test…some of the other science teachers have separate vocabulary quizzes as well

• We’ve never really gotten

       

   

132  

have the list) • Some students…have

more background knowledge coming in, so they’re able to make more connections to words they already know or things they’ve already learned about in their lives or encountered…for them, vocabulary instruction is a little easier

• If they recognize the word part, then…it’s helpful

• They already learned photosynthesis, so they could then make that connection…they could come up with ‘it’s growing toward the light’ (phototropism)

• If they saw hydrotrophic…they said ‘we’ve seen hydro before.’ They’ve seen that in ELA

• They made the connection to a hydrant, so they said, ’That has to do with water’ (hydro). They knew that the plant’s roots grow towards the water

• It definitely helps with them knowing what the word means. Not that they necessarily know the scientific process part of it, but at least they can connect that vocabulary word to that idea

what this means already?’

• If it’s something brand new, I…give them the definition for it

• If it…has a word part that I think they’ll recognize, for example when we do visible spectrum we talk about ‘what does visible mean? Do you know any other words that have visible in them?’ We’ll go from there

• I’ll usually have pictures to go with them

• Hands on activities; if it’s the parts of a plant, they’ll have those parts in front of them

• For the most part (science teachers have the same vocabulary words)

• If there’s overlap, if I know that there’s one that they’ve seen before (in ELA, I will make reference to it)

• I try to make those connections very explicit

• I’ll say ‘photo; let’s try to think of all the other words we can think of that have photo in them’

• I definitely instruct them to do that quite a

together and created a master list of vocabulary words. We don’t use common assessments in science

• Correct (there’s no formal central list of words)

• The words that they’re studying in ELA, the roots don’t necessarily come up in any of the science words that we have…there isn’t a lot of overlap

• For (some) students it’s brand new unfamiliar territory, they have no background knowledge, they might have deficits in terms of their reading abilities, and so they don’t have that vocabulary foundation, then it’s a lot more difficult for them

• They don’t really make the connection that a photograph has to do with light

• Not at this time (coordination of strategies)

• We’ve really been working more on our math curriculum and have not yet worked on coordinating science curriculum to a great degree

       

   

133  

bit (break down word parts)

• ‘Do you remember what photo means? That means light, so what do you think phototropic is?’

• Everyone (each science teacher) does have a list of the Greek and Latin roots that they do in ELA. I made a specific list of roots that related to math and science that were distributed to everybody

Second cycle coding for interviews. The four factors for determining categories for

themes as described by Krueger and Casey (2009) were also used for the second cycle coding of

the interviews. These included frequency, specificity, emotion, and extensiveness. Under these

indicators, the same five themes that emerged in the focus groups for growth/achievement and

difficulties/challenges were indicated for the interviews.

Both the ELA Curriculum Leader (ELA CL) and the science Curriculum Leader (Sci.

CL) are fifth grade teachers. Their interview responses will be analyzed under each of the five

resulting themes for each category as well as a list of the strategies they reported using for

vocabulary instruction.

Growth/Achievement  

The five themes of student growth/achievement were student motivation and

engagement, connections to other subjects, conceptual understandings, student skill

development, and teacher engagement.

       

   

134  

In terms of student motivation and engagement, the ELA CL said that her students

“love to find words that have roots in for the week…sometimes they’ll say, ‘You know what I

saw on TV…?’” and “They’re excited about it.” She continued, saying that, “They try to outdo

each other…it creates competition…some go way above and beyond, like they like it; it’s a

challenge for them.” She found that using this form of vocabulary instruction is “very

engaging…this year the kids are somewhat low overall, and they’ve just done amazing with

them (roots).” “They’ll say, ‘this is fun. That’s a good way to do this. I never thought of this

before.’…This group was really eager to make those connections and they did it on their own.”

She said the students have “built an admiration for learning new words…they look forward to

it,” “They want to build, they want more words, they want to be able to use them and show them

off,” and “They’re just so excited about it.” In class, “The kids will say, ‘Stop! That’s one of the

Greek and Latin words that we had’, (and) I always do something like, ‘Oh my goodness, I have

no idea what…flourish means,’ and they go ‘Flor, flor! That means to thrive, grow well!’.”

Finally, this CL says that this program gives the children a sense of pride. “One of our (school)

themes was ‘How does something inspire you?’ Right away they kept referring to ‘breathing it

in, breathing it all in!’”

The second theme is the connection to other subjects. The ELA CL described

connections that she observed students making to other content areas. She said, “I see them

connecting to science.” For example, students identified “biomes” when they reviewed “bio.”

This also happened with “‘Vor, to eat,’ they went through ‘omnivore, herbivore,’ they made that

connection to (the science teacher’s) class…they’ll say to me, ‘Oh, we learned that in science’.”

This also occurred in social studies. “When we talked about ‘lat’ and ‘long,’ they were like ‘Oh,

like the latitude and the longitude that we did?’” Also, when they talked about “the branches of

       

   

135  

government, the students said ‘Preamble, pre…that means before. That was the introduction to

the Constitution!’” She added, “They kind of make the connections… maybe more so in science,

but I think all the way across I see the connections.” There are connections being made outside

of school subjects as well. “There’s all these ‘ah ha’ moments where the kids will say, ‘That’s

Greek or Latin’ or, ‘That’s a Latin root’…I see home connections like ‘dorm, sleep.’ (Students

say) ‘Oh, my sister lives in a dormitory’ or, ‘My brother is in a fraternity.’ They’re making real

life connections as well.” As far as the scope of the growth/achievement, she said, “In general;

all of (the students) make those connections.”

The Sci. CL said that the students coming in with more background knowledge are “able

to make more connections to words they already know or things they’ve already learned about in

their lives or encountered…for them, vocabulary instruction is a little easier.” She also gave

examples of connections that students were able to make to other words such as

“photosynthesis…they could come up with ‘it’s growing toward the light’ (phototropism).”

Both CLs also commented on topics related to the third theme, conceptual

understandings as well. The ELA CL said, “They have all these ‘ah ha moments’.” She has

observed that, “They remember what the words mean…It’s filled their vocabulary with words

that were provided as well, and they’re taking when words are out of context they’re able to put

the words together by using the roots.” She added, “I can’t stress enough how much they

continue to use those words of the week; they’re now embedded in their

vocabulary…Absolutely. Yep, absolutely (it helps them understand concepts).” The Sci. CL said

that studying the word part “definitely helps with them knowing what the word means. Not that

they necessarily know the scientific process part of it, but at least they can connect that

vocabulary word to that idea.” For example, when reviewing ‘hydro,’ “They made the

       

   

136  

connection to a hydrant, so they said, ‘That has to do with water.’ They knew that the plant’s

roots grow towards the water.”

Student skill development is the fourth theme. According to the ELA CL, “They

remember what the words mean…they’re taking when words are out of context they’re able to

put the words together by using the roots.” She said that they use these skills to hypothesize the

meanings of unknown words “all the time.” Even “when there’s no context clues at all…they’re

uncovering new meanings…it’s not always accurate, but they attack it.” The students tell her “‘I

never thought of this (strategy) before.’ They have built and they’ve acquired new words, but I

think they’ve built an admiration for learning new words.”

Then final theme in this category is teacher engagement. In terms of ELA, the CL said,

“I think it’s a really great program.” The Sci. CL gave no clear indication of her level of

engagement with the program.

Difficulties/Challenges  

Again, there were also five themes identified under the difficulties and challenges

comments made in the interviews. These themes were lower achieving students, program

coordination, time/resources, level of mastery, and teacher mindsets/capacity.

The ELA CL specifically addressed lower achieving students. She said, “The low, low

kids have trouble making connections…to trying to find roots on their own and be able to apply

them.” She added, “Overall the students are so low that they’re not making that connection form

the year before.” In science, the CL said that “For (some) students it’s brand new unfamiliar

territory, they have no background knowledge, they might have deficits in terms of their reading

       

   

137  

abilities, and so they don’t have that vocabulary foundation, then it’s a lot more difficult for

them.”

The second theme captures the program coordination. The ELA CL will ask students

“‘didn’t you do this last year?’ ‘No,’ some of them said…’” In terms of coordination, the science

CL said, “We’ve never really gotten together and created a master list of vocabulary words. We

don’t use common assessments in science.” She said there is not a formal central list of words

that all teachers share and that there are many roots that they are studying in ELA that don’t

necessarily come up in science. She said that there is also not a current shared pool of strategies.

In terms of curriculum, she said, “We’ve really been working more on our math curriculum and

have not yet worked on coordinating science curriculum to a great degree.”

The ELA CL did not clearly make any comments that reflected on time and/or

resources. In science, the CL said they are not yet using the Next Generation Science Standards.

She made no other direct references to resources or time.

The fourth theme, level of mastery, was discussed only briefly. When referring to the

majority of students (and not the challenged learners), the ELA CL did not share any negative

observations regarding student level of mastery. The Sci. CL provided one example of difficulty

with mastery when she said, “They don’t really make the connection that a photograph has to do

with light.”

The final theme is teacher mindset/capacity. There was nothing in the ELA CL

comments that were appropriate for this category. The Sci. CL simply provided an explanation of

the state of the science curriculum. She said, “We’ve really been working more on our math

curriculum and have not yet worked on coordinating science curriculum to a great degree.”

       

   

138  

Fidelity/Strategies  

The final column discussed specific strategies in teaching vocabulary. In English class,

the CL would “present the vocabulary; I’d have them look at the board to see what they thought

they saw in common until we came to the root word, and then I would tell them the…Greek and

Latin words that they came from and we’d talk about the definition.” She would also have them

“come up with words that (they) knew…look them up; do some application of it…They had to

find several examples of other words that they found the root in, they wrote sentences with them;

they wrote stories with them.” She also “discussed them indirectly oftentimes with reading

selections.” She would model lessons and desired student thinking and made notes of the science

terms that had Greek and Latin word parts in them. Other strategies included “find(ing) the

affixes…weekly” which she described as “very helpful.” She assigns ten words per a week and

there’s “a set time during the day…every single day, that (they) talk about (vocabulary).” She

feels that high student performance “all depends on the fidelity of the program.”

The Sci. CL introduces the vocabulary words with PowerPoint. She displays the word

and said, “We might talk about it ahead of time. If I think they’re already familiar with the word

I might ask them, ‘Do you know what this means already?’ If it’s something brand new, I…give

them the definition for it.” She will identify the word part if she thinks they are familiar with it.

For example, she said, “when we do visible spectrum we talk about ‘what does visible mean? Do

you know any other words that have visible in them?’ We’ll go from there.” She utilizes visuals

and hands on activities. “I’ll usually have pictures to go with them…if it’s the parts of a plant,

they’ll have those parts in front of them.” She makes specific reference to students any words

that she knows are also studied in their English class. She said, “I try to make those connections

very explicit. I’ll say ‘photo; let’s try to think of all the other words we can think of that have

       

   

139  

photo in them.’ I definitely instruct them to do that quite a bit (to break down word parts).” She

shared an example where she said to students, “‘Do you remember what photo means? That

means light, so what do you think phototropic is?’”

In terms of fidelity, she said that “for the most part,” all of the science teachers have the

same science vocabulary lists. “Everyone (each science teacher) does have a list of the Greek and

Latin roots that they do in ELA…(she) made a specific list of roots that related to math and

science that were distributed to everybody.”

       

   

140  

Chapter V: Discussion of Research Findings  

“The limits of my language mean the limits of my world.” --Ludwig Wittgenstein

Introduction

The purpose of this study was to determine if students were able to apply what they have

learned in the study of morphology in their ELA classes to their understanding of content area

vocabulary in their science classes. A further goal was to identify whether students were able to

use any conceptual understanding of Greek and Latin word parts to hypothesize the meanings of

nonsense words that included science-related morphemes.

It has been widely reported and documented that there is a strong connection between a

robust vocabulary and reading comprehension (Chall and Jacobs, 2003; Beck, McKeown, and

Omanson, 1987; Stahl and Fairbanks, 1986; Graves, 1986; Nagy, 1988; Anderson and Freebody,

1981; NICHHD, 2000; Sinatra, Berg, and Dunn, 1985; Baumann, Kame’enui, and Ash, 2003;

Hiebert and Kamil, 2005; Mezynski, 1983). William Nagy (2005) also suggested that there is a

bidirectional relationship between vocabulary and comprehension, which he refers to as the

Reciprocal Model. It is also understood that vocabulary is an essential “strand” in the

comprehensive “reading rope” of children (Scarborough, 2001). When children have not

mastered grade level reading proficiency by the intermediate grades (approximately grades four

and five), they may suffer from the “Matthew Effect” (Stanovich, 1986). Each of these theories

underscores the necessary sense of urgency in making sure that every child is able to read with

comprehension at grade level by the time they enter middle school. If not, the statistics for failure

are staggering.

       

   

141  

According to reports by Achieve, Inc. (2005) and the Carnegie Corporation (Biancarosa

and Snow, 2004), students are not adequately prepared for college or career upon leaving high

school. This has resulted in an inordinate number of remedial courses needed for struggling

college freshmen and employers who are not satisfied with the level of basic skills of the entry

level work force. In addition, recently released Common Core State Standards for ELA and

Mathematics as well as the Next Generation Science Standards all call for a heavy reliance on

effective communication including scientific literacy.

Science, in particular, is an area in which students have reported weaknesses. In the

Achieve, Inc. study (2005) of high school students, “51% felt there were gaps in their preparation

in science.” There are three distinct contributors to the difficulty of accessing and understanding

science vocabulary. First, the conceptual nature of the domain specific vocabulary can be

problematic. Unlike other content areas, the introduction of a science term is likely also the

introduction of a concept that is entirely new for the student. This makes it difficult to compare it

to any other word the child may know. Having a conceptual understanding of part of the word,

however, could help the student to get an idea of what the word represents as they learn the new

application and concept. For example, if a new word has “hydro” in it, the child would recognize

that the word has something to do with water or fluid.

Second, the technical subjects, like the sciences, are utilizing a vocabulary that is growing

exponentially with continual inventions, discoveries, and innovations. Learning even the most

common affixes and base words can help a student to learn far more words and access a greater

number of concepts. This is particularly true in the sciences where most scientific terms are

based on Greek or Latin roots.

       

   

142  

Finally, direct and explicit vocabulary instruction has not traditionally been part of

science lessons. This is typically done in the English classes without much coordination or

articulation between teachers or courses. Capitalizing on students’ experiences across classrooms

could help in creating repetition of exposure as well as the use of quality vocabulary acquisition

strategies.

For this study, a total of 397 students (201 in grade four and 196 in grade five) and their

teachers participated. Results of a teacher assessment, two researcher-created student

assessments, four focus groups, and two individual interviews were presented. As a result, each

of the research questions can be addressed. Again, these questions were:

1. A. How do students apply their understanding of Greek and Latin word parts

in ELA to acquiring vocabulary and understanding new concepts in science?

B. What contributes to student application of morphological strategies learned in ELA to

their study of vocabulary in science?

2. A. To what extent do students utilize their understanding of morphemes to decode and

comprehend or conceptualize unfamiliar scientific vocabulary?

B. How do students hypothesize the meaning of nonsense scientific vocabulary

utilizing their understanding of Greek and Latin word parts?

Findings

The lack of coordination of the program must first be discussed in order to provide an

accurate context for the findings. At the beginning of the study, the building administrator and

both of the Curriculum Leaders explained that the Prestwick House scope and sequence and

workbooks were being used by both grade four and grade five in all ELA classes. The ELA CL

       

   

143  

provided a list of the vocabulary words and indicated that classes were at approximately the

same unit in the series. The Science CL provided a list of the units taught in science with all

relevant vocabulary. She further explained that every science teacher had been given a copy of

the ELA words being taught in each grade level. It was the presumption of the three leaders that

all grade four and grade five teachers were utilizing the program and coordinating their lists of

vocabulary words that had Greek and Latin parts. However, as a result of the teacher focus

groups, it has been determined that this was not the case.

This misunderstanding regarding the program coordination on the part of the instructional

leaders is likely due to the fact that both Curriculum Leaders are fifth grade teachers and they

report directly to the principal. They work most closely with their fifth grade colleagues and have

common planning with them. This allows them to directly observe what is happening in the fifth

grade classes. There is not a formal system of observing the implementation of any curriculum or

program in other classes. If it is not reported by teachers that they are not following the program,

it is not understood to be the case. There is a presumption that it is being done when it is not.

Also, while the district had begun a process of creating individual units of study utilizing the

Understanding by Design process, they have not completed these units. Common assessments

and resources have not been clearly chosen or used consistently across grades or subjects.

Therefore, if a teacher does not seek out his or her own “supplemental resources,” they do not

have what they need to advance the program, and it is not likely that it is consistent across a

grade or aligned with other grades.

This unexpected development created a research opportunity, however. Now, an added

feature could be disaggregated from the data. It would now become possible to analyze whether

the use of direct, explicit instruction of the Greek and Latin word parts had a different effect on

       

   

144  

the achievement of the students in their science classes compared to a group that did not

emphasize the identification and reflection on the morphology of the words.

Finding A: Fidelity and collaboration

According to the grade four ELA teachers, they are not closely following the program. As

noted earlier, they find that it is “hard to squeeze in with other things” and “with time, it’s just

not feasible.” They do not employ various strategies saying that when they do use the resources,

they “follow the workbook, page-by-page” and do not complete the activities “every week.”

They also acknowledge that there is “no collaboration” with their science colleagues in regard to

vocabulary. They complained of a lack of resources, but they did not specify what resources they

believe they needed. The grade four science teachers agreed that when it comes to vocabulary,

there is “no collaboration.” They, too, were not happy with their materials, saying that they were

not aligned to the new Massachusetts standards and that they were “a million years old.” They

said that their current “textbook doesn’t make connections” to the Greek and Latin word parts for

them. While they used more different kinds of activities in teaching vocabulary than the ELA

teachers described, they were not done with any reference to the Greek and Latin relationships.

They concluded, “We are going to have to do something different soon,” but added, “what that is

yet, (we) don’t know.”

While the grade five science team said they had “not too much collaboration” on the

vocabulary with their ELA partners, they said they had gotten the list of words covered in ELA.

Two teachers pointed out that they go into the ELA classroom when they notice a word that is

being covered in both classes. One of them “will go put it on her list (of Greek and Latin words)

so the student will see it in both places.” The other “will put it on her board so (she) sees it or the

       

   

145  

students will see the same word in my class and their class.” While they also said they had a lack

of resources, they said that they secured “supplemental resources” on their own.

The researcher-created, multiple-choice vocabulary assessment challenged students to

apply their understanding of Greek and Latin-based words they had learned in ELA, science, and

both ELA and science. The tool measured students’ understanding of words that should have

been familiar to them. This directly related to the first research question:

How do students apply their understanding of Greek and Latin word parts in ELA to acquiring

vocabulary and understanding new concepts in science?

The students in each grade performed well on words that they had learned in their ELA

classes. This included words that were repeated in the science curricula. In fact, they performed

identically in both grades. Students in grades four and five averaged 81% for ELA words and

83% for words from both ELA and science. However, for words that were only covered as part

of the science curricula, the fourth grade students averaged 67% correct, while the fifth graders

averaged 83% correct.

While grade four and grade five students had varied experiences with the Greek and Latin

vocabulary in their ELA classes according to the teachers, they performed well in the assessment

of words they had been exposed to that contained Greek and Latin word parts. However, the

grade four students did not have this same level of retention with words that they had only

encountered in their science classes. This would indicate that the more direct, explicit instruction

and efforts to make cross-curricular connections clear to students in grade five ELA resulted in

higher student achievement. Teaching students to transfer this strategic approach to

deconstructing scientific vocabulary into manageable word parts helped them to more accurately

define terms in science class.

       

   

146  

It was through the focus groups and individual interviews that the level of program

fidelity became clear. This had a marked impact on the analysis of the resulting data. In addition

to the student assessment results, many of the teacher responses also correlated with the level of

program implementation by grade. The highest number of strategies and instances of teacher

collaboration were reported in grade five, and this is the grade that had the highest level of

science vocabulary success.

Since the teachers in this school already have the Prestwick House resources, it would be

relatively easy to coordinate their implementation of the scope and sequence of these word parts

across content areas. For example, even a cursory review of the word parts (roots only) taught in

several of the Prestwick House ELA units for each grade include the following potential

relationships to other content areas:

Subject Mathematics Social Studies Performing and Visual Arts

ELA

GRADE 4 metr, centi- frag, fract, tract circ, cycl prim, quart, second, oct

temp, ann, chron circ dict, voc mag, min amb cap, manu

aud, vid, vis dict, voc prim, quart, second, oct son, phon

graph scrib gram bibl, verb, leg, lec, log ling

GRADE 5 uni, duo, tri init long, lat cur, curs mim, siml integr fin, term morph, form vers, vert angl, gon, rect sum, cumul, neg

mater, matr, pater, part, frater urb, loc, poli uni, duo, orig, arch domin, dom vict, vinc, reg, dyn long, lat doc, drom, cur aer, lith, agri migr, tort volv, volute cred, dox, fid

uni, duo, tri init mim, siml morph, form vers, vert angl, gon, rect vac, neg

vict, vinc reg dyn doc, mon drom cur, curs mim, siml sum cred, dox, fid

       

   

147  

Finding B: Strategic approaches

While the grade four ELA teachers reported varying degrees of fidelity to the program,

they all “covered” the words provided in the workbook. The grade five ELA teachers indicated

using a greater number of strategies to approach the vocabulary instruction including many that

were metacognitive in nature. As reported by the science teachers, the grade four teachers did not

make connections to Greek and Latin or to what the ELA teachers were doing in their classes.

However, the fifth grade teachers did make these connections. All of the fifth grade science

teachers also made use of word attack strategies in their vocabulary instruction, or “breaking

down” words to their component parts when determining their meanings. While simple recall

may account for successful student performance in ELA, the strategies that best served students

in applying their understanding across content areas were those that were more reflective and

strategic. These are the skills they learned in the fifth grade ELA and science classes and this

most directly addresses the following research question:

What contributes to student application of morphological strategies learned in ELA to their

study of vocabulary in science?

Grade five teachers described far more activities in teaching vocabulary that are above

and beyond what are in the workbooks. These include activating prior knowledge and thinking,

writing and drawing with the words, playing games, and using graphic organizers. Each activity

references the Greek and Latin connections. These represent differentiation and higher levels of

cognitive skills. Subsequently, their observations of their students’ vocabulary skills included far

more references to “student connections” than in the fourth grade. Grade five teachers referenced

positive “connections” 49 times as opposed to six times in grade four.

       

   

148  

The science teachers acknowledged different degrees of emphasis on Greek and Latin

instruction. While half said they pointed out the Greek and Latin roots to students when they

came across them, the other half said they did not. However, all of them said that they utilize

word attack strategies or have students “break down the words” into their smaller parts to

identify their meanings. In fact, this strategy was observed in the nonsense word assessment and

was referred to as the “separation strategy.” This included students underlining the compound

word parts, making a slash between the parts, or circling the two parts of the words. While both

fourth and fifth graders did this, the fifth grade students did it most successfully. When this

strategy was employed, 71% percent of the fifth graders identified the concept correctly as

opposed to a success rate of 46% for the fourth graders.

One fifth-grade teacher presented assessment results from the Prestwick House series.

This longitudinal archival data was made up of pre-test and post-test scores for two units of the

Greek and Latin program. Each unit was made up of five lessons. This represented

approximately half of the school year as each lesson is done in about two weeks. The same tool

was used for the pre-test and post-test. This tool was the five-lesson cumulative review provided

in the workbook. It was not surprising that the students performed better on the post-tests after

two weeks of instruction. They scored an average of 80% on these assessments (80.5% and

80.6% respectively). However, the pre-test scores increased from the first unit administration to

the second. Students scored 9% better on the second pre-test (53.3% to 62.3%). This could

indicate that their word attack strategies had improved after a complete unit of study and that

their approach to unknown words was different. Collection of the two remaining units would be

necessary in order to know if this is valid. Having more data points would help to confirm this

speculation.

       

   

149  

Finding C: Conceptual understandings   The second researcher assessment was created using nonsense words that contained a

Greek or Latin word part that students had learned in ELA, science, or both ELA and science.

Using a “nonsense” word eliminated the possibility that students could define the word based on

word recall alone. They would have to apply a conceptual understanding of the root. This

assessment would be far more indicative of student mastery of understanding and, perhaps more

importantly, their transference of this understanding to other contexts. This directly addressed the

second research question:

To what extent do students utilize their understanding of morphemes to decode and comprehend

or conceptualize unfamiliar scientific vocabulary? and How do students hypothesize the meaning

of nonsense scientific vocabulary utilizing their understanding of Greek and Latin word parts?

While students scored similarly in identifying word parts they had learned in both ELA

and science (58% correct in grade four and 50% in grade five), students in grade five

outperformed students in grade four in each individual content area. Students in grade five were

able to correctly identify the concepts of word parts learned in ELA 73% of the time (as

compared to 42% of fourth graders). The fifth grade students could also correctly identify the

meanings of 54% of the word parts learned in their science classes as opposed to fourth graders

who could only do this 35.6% of the time.

Overall, the highest performance was measured in fifth grade ELA, which correlates

perfectly with the amount of direct, explicit instruction that was given in relation to the Greek

and Latin word parts in these classes. The second highest performance was in fifth grade science,

which had the second highest reported level of strategic and purposeful instruction.

       

   

150  

Another indicator of a higher level of understanding among the fifth graders was actually

in the errors made on the nonsense word assessment. Two of the three most common errors

among fourth graders were the nonsense words “exorain” and “terrahouse.” The most common

answers were those that referenced words that had nothing to do with the conceptual meaning of

the root word. For example, student responses had to do with “x-rays,” or “exercise machines”

for “exo,” and multiple variations of “terror” for “terra.”

Conversely, two of the three most common errors for grade five were confusing the root

with other words that did have relation to the correct meaning. The nonsense words with the

concepts that were misidentified were “matricake,” and “omnicoors.” Several students identified

“matricake” as having to do with “family,” “wedding,” or another specific family member.

Others correctly extracted the “tri” from the word and connected the meaning “three” to the

concept. While this was counted as an incorrect response, they did correctly identify a word part

that they had learned this year.

Many fifth grade students (35) also incorrectly identified another nonsense word,

“omnicolors.” However, the majority of these responses had to do with omnivores. While

students correctly connected the root to the word omnivore, they incorrectly identified the

definition. They did provide answers such as “colors that eat both meat and plants” or “meat and

plant colors.” This could be the result of the definition they were taught by teachers. In fact, the

science curriculum provided identified an omnivore as “organism that eats both plants and

animals” vs. “organism that eats all things.” Therefore, the students were applying the concept

correctly based on what they learned. An explicit explanation of what “omni” meant in reference

to an “omnivore” would have helped students to have a better understanding of the meaning of

“omni” for future reference and application.

       

   

151  

Students had never encountered these words before because they were, in fact, nonsense

or made-up words. However, those students that had previously learned the concepts behind the

word parts were better able to understand the potential concepts behind the foreign words when

presented in a different content area because of that foundational understanding. This supports

the research of Tomlinson (2010), when she states that, “When students understand, then we find

that they can retain things better, they can retrieve it from memory better, they can transfer it

better, they can use it in the world more effectively.”

Finding D: Student and teacher engagement

The grade five teachers, in both ELA and science, reported far more instances of student

“engagement” when utilizing Greek and Latin as a means of teaching vocabulary. In fact, the

level of student “engagement” through the use of this approach was mentioned 22 times by the

fifth grade teachers, but only three times by grade four. The positivity of the comments as well as

the body language and level of excitement of the grade five teachers observed by the researcher

were much higher and more frequent than those of the fourth grade teams. Again, this level of

engagement correlated with the level of fidelity, the use of varied strategies, and the teacher

coordination as exhibited by the grade five teachers. Indeed, their students also had a higher level

of achievement in all measures. In addition, the teachers reported a greater number of student

connections to “life” as illustrated by the their reflection of student examples of identifying

Greek and Latin parts on television, in their pleasure reading materials, in classroom texts, and in

conversation (ie: “my sister lives in a dormitory,” “my brother is in a fraternity,” and references

to a teacher with “urban” in her last name as “Mrs. In-The-City”). These examples may illustrate

that the “rich” are “getting richer” as theorized by the Mathew Effect (Stanovich, 1986). It is also

       

   

152  

an indication that as the students increased their vocabulary, they increased their comprehension,

and as they increased their comprehension and continued reading, they increased their

vocabulary. This exemplifies the Reciprocal Model (Nagy, 2005).

Other research has supported this claim about the importance of student engagement.

According to Carol Ann Tomlinson (2010), if teachers have “a really dynamic curriculum and

differentiate that, then the power of the curriculum gives more power to the differentiation, not to

mention more power to the students…(teachers must) capture their minds…evoke some

curiosity…make them wonder about something or connect with it. Engagement is one really

critical attribute of quality curriculum.”

It is interesting to note that the theoretical frameworks of this study seemed to extend to

the teachers as well as the students. Teachers in both grades reported feeling “overwhelmed” by

the program when it began and claimed to have a lack of resources and time to complete all that

is required of them. However, when it came to the vocabulary instruction, the grade five teachers

indicated they have moved past these difficulties saying that they are “more comfortable” now

and that they found their own “supplemental resources” to use that were not necessarily provided

by the school or district. The grade four teachers expressed more of a sense of helplessness when

they agreed (as stated earlier) that “We are going to have to do something different soon,” but

“what that is yet, (we) don’t know.”

The teachers also agreed that, “there is no official common planning time,” so “time-

wise, it is still very difficult to get in.” However, the grade five ELA teachers did note that while

they “initially” found this work “intimidating,” they are now “much more comfortable” with it.

As a result, one teacher said, “the more comfort I see with it, the more I see (the students’)

comfort as well.” They said that they complete the Prestwick House units and that they “know

       

   

153  

the units (the science teachers) teach and the concepts they cover” in science class. This helps

them to make vocabulary connections for their students. The more enthusiasm they reported on

the part of their students, the more enthusiastic they got in their explanation. It was as if the

“rich” were “getting richer,” and they were more dedicated to making it work, even if they had to

carve out their own time and secure their own resources.

It is the researcher’s belief that sharing the data results of this study with the fifth grade

teachers would provide positive reinforcement and perhaps continue the cycle of the Reciprocal

Model (Nagy, 2005). As teachers continue to find excitement in their students’ performance,

they may continue to elicit excitement in their students. This would cause both groups to focus

on the strategies and skills that are currently bringing them success and continue to improve the

level of self-efficacy they are currently feeling. One area that this is lacking is in the grade five

teachers’ expression that a group of students are not retaining their understandings of the word

parts. They only have one measure of this, and that is the Prestwick House multiple-choice

cumulative assessment on the words learned that include the word parts. However, the nonsense

word assessment scores collected by the researcher for grade five showed a great level of

retention and application. Their students were able to transfer their understanding 54% of the

time (as opposed to 35.6% in grade four). In ELA, where there was more emphasis on the

morphology, they were able to apply their understanding 73% of the time. This is the type of

data and feedback that allows this cycle of engagement and confidence to continue.

Implications for practice   Learning science vocabulary presents three problems that make the acquisition of

domain-specific words more challenging than in an ELA class. First is the conceptual nature of

       

   

154  

learning new words. Most often, the new word represents a novel idea or understanding for the

child. It is not a word that can be made synonymous with something the student already

understands. In this study, children demonstrated that they were more successful at identifying

the concepts associated with unknown (nonsense) science terms if they had a prior conceptual

understanding of the root of the word, even if they learned that root in ELA. This concurs with

previous research (Aaron, Joshi, and Quatroche, 2008). In fact, students were more successful at

this the more they were engaged with the learning of the word part and the more that the

connections were made explicitly to them (as in grade five vs. grade four).

The second difficulty is the sheer number of new vocabulary words required in content

areas like science. Despite the fact that it is estimated children will need to learn up to 3,000

words per year to maintain grade level vocabulary, they can only master approximately 400 per

year through direct instruction (Beck, McKeown, Kucan, 2002). Science vocabulary is largely

based on Greek and Latin derivatives. In fact, 90% of multisyllabic words are based on Latin,

with most of the remaining 10% based on Greek (Rasinski, 2008). It is estimated that a student

can generate five to twenty different words from one Latin root (Rasinski, 2008). Therefore, if

the students of Billings Intermediate School were to study and learn just four root words per

week, they would have the potential of having a conceptual understanding and be able to

successfully hypothesize the meaning of 800 to 3,200 words per year, many of them highly

technical and domain-specific. According to Beck, McKeown, and Kucan (2002), students

generally have the capacity to learn 8-10 words per week through direct instruction.

The third challenge to learning vocabulary is lack of experience most teachers (other than

ELA teachers) have in language-based literacy strategies such as teaching vocabulary. However,

the newly released Common Core State Standards, Massachusetts Curriculum Frameworks, and

       

   

155  

Next Generation Science Standards all place a heavy emphasis on literacy skills such as

“understanding words and phrases, their relationships, and their nuances, and on acquiring new

vocabulary, primarily general academic and domain-specific words and phrases” (MA DESE,

2012). This will require direct and explicit instruction of all teachers in collaboration and

cooperation that must go beyond mere memorization to conceptual understanding and proper

application. Researchers have not conclusively determined one strategy that is most effective for

all students. This speaks to the fact that differentiation is key. In fact, this study has reinforced

that understanding by illustrating that the most successful students (grade five) were those whose

teachers utilized many and various strategies in their teaching of vocabulary, including those

strategies that required higher levels of cognitive demand, including metacognitive reflection of

their background knowledge and of the strategies themselves.

Throughout the study, teachers in both grades also made reference to the “connections”

they observed students making to other content areas. These included mathematics and social

studies, as well as “TV,” “other readings,” and “life.” A potential implication of the results of

this study would be the replication of the success of the fifth grade students’ vocabulary

acquisition in these other areas as well. This would be particularly true with some modifications

to the current program and how it is implemented as noted in the recommendations for

educational practice below.

Recommendations   The recommendations from this study come most directly from the themes identified in

the focus group and interview processes. Even though these conversations took place after the

       

   

156  

student assessments were given, the data from the assessments substantiated each of the themes

and indicated which were more successful and which were not in regard to student achievement.

Encourage student motivation and engagement. Teachers in this study noted that

students responded best when they viewed the activities as a “competition” or as a means to

“challenge” themselves. They “want to build, they want more words, they want to be able to use

them and show them off…they’re just so excited about it.” This has helped to provide a much

deeper level of understanding of the terms. “As a result…they have built and they’ve acquired

new words, but…they’ve built an admiration for learning new words…they look forward to it.”

Again, grade five teachers reported 22 times during their focus groups and interviews that their

students found this work to be engaging as opposed to three mentions of interest by grade four

teachers. Subsequently, the grade five students outperformed the grade four students in the

assessments.

Make cross-curricular connections. Teachers in both grade four and grade five noted

that students were “making connections” to words used in other classes. There were many more

references to connections in grade five where the strategy of using words across content areas

was more deliberate (49 references vs. six in grade four). This would be a relatively simple

addition to the curriculum at the Billings Intermediate School. This is because the current scope,

sequence, and materials being used (Prestwick House) have a collection of words that could

easily be added to the list of words taught in other content areas. They are already being taught in

the ELA classes. Connecting the words to what they may already be learning in other areas

would be a matter of proper planning and communication.

Focus on conceptual understandings and differentiate instruction. Most subject area

content, including science, is best understood conceptually. Students must connect new

       

   

157  

information to what they already know in order to make sense of it and be able to apply it

accurately. Teaching vocabulary words in isolation as part of a random list of words is not

effective. It is also not practical with the number of words they must learn and their capacity for

learning only a small fraction of that from direct instruction.

The classes with the highest levels of achievement in this study were the ones that

incorporated activities that were varied and required higher levels of cognitive engagement. The

grade four teachers noted seven activities that they used in their classes when teaching

vocabulary. The grade five teachers noted no less than 27 strategies. Also, while the grade four

activities consisted of going through “the workbook page-by-page” and utilizing “flashcards,”

the fifth grade teachers encouraged students to “activate their thinking,” “find (other) words”

utilizing the same root(s), identifying where the roots are in the words, and completing Frayer

Model graphic organizers which are a better indicator of student conceptual understanding.

Utilizing multiple strategies also encourages individual student differentiation based on their

needs and learning styles.

Cultivate teacher engagement through professional development. Teachers in both

grades expressed being “overwhelmed” and “intimidated” by the program at some point.

However, the fifth grade teachers reported that, with increased practice, they became “more

comfortable.” As a result, the students became “more comfortable” as well. Some of the fourth

grade teachers also explained that they did not know which words contained roots or whether or

not they were Greek or Latin. They said that their textbooks did not “make that connection” for

them. This resulted in uncertainty for them and the feeling that they have a “lack of resources.”

This affected their level as self-efficacy as they reported far less confidence in using the

program. Only one teacher in fourth grade said she “liked” the program, but quickly qualified her

       

   

158  

statement by saying, “only because I took Latin for a few years…I get the correlation.” Fifth

grade teachers, however, in addition to describing their “comfort” level added that they feel it is

“a really great program.” There were six positive comments by the grade five teachers as

opposed to the one in grade four. This was reflected in the level of fidelity that was committed at

each grade. However, not having the appropriate training and resources is daunting, frustrating,

and simply unfair. Having a clear set of expectations and direction would support the teachers

more appropriately.

Create a coordinated program with oversight and accountability. A coordinated

program is critical when trying to make explicit connections for students. It cannot be expected

that children will see these connections on their own. In spite of this, however, several teachers

reported just that, saying that students “were excited to inform the teacher that they knew all

about the word before she even spoke about it,” they “were noticing it before (teachers) were

instructing it,” and they “would make that connection without doing a structured Greek and

Latin…lesson.” Students often reported that they were learning a particular word part in math or

said, “Oh, we learned that in science!” A structured and coordinated scope and sequence across

content areas would help facilitate these “connections.”

While the Prestwick House series offers one possibility, it is not the only option. Simply

meeting and aligning the concepts in each area would produce the vocabulary that is common

across content areas. It is preferable to have the vocabulary result from the purposeful curriculum

that is creating utilizing the Understanding by Design process and aligned with the

Massachusetts Curriculum Frameworks. This study has presented evidence that multiple and

diverse activities are more effective, so utilizing these strategies with the identified words would

be fine as long as they are clearly articulated and well understood by the teachers.

       

   

159  

In order to insure that the program is being implemented effectively, there has to be some

oversight of the program. Even with the best intentions of implementation, a teacher without the

proper training or supports will have difficulty with any program. There must be clear goals and

outcomes for student performance. This understanding should be calibrated among all

stakeholders including teachers, administrators, parents, and students to insure success.

Provide time and resources. In order to create an aligned and coordinated program,

provide professional development, encourage teacher efficacy, and foster effective

communication among teachers and across grades, there must be ample time and opportunity for

collaboration. It is difficult to create a regularly scheduled common planning time, but the

success of any new program depends on this focused time together. Teachers will not have an

investment in the success of the program if they are not involved in the decision-making

regarding its integration. They will also not have a full understanding of the choices made for its

use or how it should be properly implemented. They should also participate in creating the

assessments to be used. They should define what the indicators of success would be and then

analyze the resulting student data to see if instructional adjustments need to be made and what

those adjustments might be.

It may or may not be necessary to purchase additional outside resources for this

endeavor. However, the units of instruction should be completed to identify the concepts and

skills to be taught in each unit. This will facilitate the choices for vocabulary and subsequent

resource needs, if necessary.

Further Research  

       

   

160  

There were questions that arose that were beyond the scope of this study. In addition,

shortcomings of this research presented themselves once the study was underway. Each of these

provides opportunities for further research on this topic.

ELL and special education. There were zero students in the Billings Intermediate

School that were identified as English Language Learners. In addition, the students that did

participate were not identified as having an Individualized Learning Plan (IEP) or learning

disability. Therefore, these variables could not be taken into account when analyzing student

results. However, teachers did comment that they felt their “low” students had most difficulty

with Greek and Latin word parts. It is unclear whether or not this was in reference to ELL or

special education students. Further study with these populations would be necessary to determine

if these students displayed different levels of achievement on these assessments.

Increased longitudinal data. The only archival data that were collected were two, five-

lesson units for grade five. This data indicated that both post-tests showed an increase in student

understanding form the pre-tests. However, the second pre-test also showed a 9% growth from

the first pre-test. This may indicate that the students were becoming more proficient at using

word attack strategies or breaking down word parts for meaning. However, without further data

points, this is non-conclusive. There could be four of these data points in a school year under the

current model at this school. Consistent growth in all four would be recommended before

making such a claim.

Comparative fidelity across participating grades. It was the intention of this research

to study the comparison across grades implementing the same program with the same level of

fidelity. However, after it was discovered that this was not the case, it became a study of

       

   

161  

contrasts. Another researcher may carry out the original goal of measuring the effectiveness of

this strategy across multiple grade levels.

Classroom observations. Classroom observations were not a part of this study.

However, this could be a valuable addition. Observations can be used to calibrate the information

and data collected by other means including the student assessments, focus groups, and

individual interviews. The level of explicit instruction and fidelity to the Greek and Latin scope

and sequence could be compared to the teachers’ self-assessments of these areas for accuracy in

reporting. The purpose of using this approach would be to “obtain data about behavior through

direct contact and in terms of specific situations in which the distortion that results from the

investigator’s being an outside agent is reduced to a minimum” (Kluckhohn, 1940).

Utilizing this model will also allow an objective view of the teacher/student interactions,

which may remove any bias, even unintentional bias, in a teacher’s report. According to the

constructivist-interpretivist construct, “lived experiences may be outside the immediate

awareness of the individual” (Ponterotto, 2005). The daily demands on the teacher in the

classroom is such that they may not be aware of the minutiae that might indicate student

recognition or may serve as an example of particular insight or confusion on the part of a child.

Observation for these specific elements and examples may uncover these details that may

otherwise go unnoticed. This would give a greater level of confidence in the accuracy of

resulting student performance data.

Student follow-up interviews. There were no student interviews conducted as part of

this study. However, upon review of the student assessments, particularly the nonsense word

assessments, there were several opportunities for follow-up questions with students. In the event

that a picture was indiscernible or the wording of a definition was unclear, the item was marked

       

   

162  

incorrect. However, it would have been interesting and possibly insightful to have the students

further explain their reasoning and representations. Also, there were many teacher reports of

“student engagement.” It would have been helpful to survey students themselves on their level of

interest in working with words in this manner. Measuring student interest in the context of their

achievement may have provided further useful data.

Other content areas. The morphological approach to content area literacy may have the

same results in other content areas. Theoretically, this would be the case. It may also be possible

to see even higher results if the vocabulary words were across three or more content areas and

not just two. The level of student conceptual understanding could then be measured in other

ways as well. For example, if a word has more than one meaning depending on the content area

in which it is studied and applied, it would be interesting to note whether the child is able to

discern the uses across disciplines. This would show a more sophisticated level of understanding

and application.

Fluency and spelling skills. It was reported by one teacher that the Greek and Latin

word study helped her students with “syllabication.” Upon further clarification, she confirmed

that students were better able to pronounce the words due to their “chunking” of the individual

word parts. Further study could be done to examine whether morphological study helps to

improve student reading fluency. It could also be investigated whether or not it improves student

spelling proficiency as well since they may memorize letter combinations that are less familiar to

them such as “sym,” “photo,” “amphi,” or “terra.”

Program review/case study. A specific case study could be completed for a particular

school and/or program. This could include a published program, a particular scope and sequence,

a chosen curriculum unit(s), or even a set of strategies used to teach the word parts. This could

       

   

163  

also be used to measure the success of an implementation, a professional development initiative,

or common planning time across a grade or across particular content areas.

Conclusion   Strategies for vocabulary acquisition in the content areas have been largely overlooked in

literacy literature and research. Recommendations have been made to simply apply the

instructional practices that have been used in language arts classes. However, the unique nature

of the content areas and their domain-specific vocabulary have made this difficult and frustrating

for teachers who do not identify themselves as literacy or reading teachers. However, state and

national expectations for college and career readiness have made the definition of a literate

graduate more demanding than ever before. Accessing dense expositional text independently is

an expectation of every student. Having the unique characteristics of the content areas in mind

when learning the pertinent vocabulary such as the conceptual nature of the terms, the massive

quantity of them that must be learned, and the need for direct, explicit skills to be used by all

teachers, can result in an effective and successful program. Further, encouraging a culture of

teacher support, professional development, collaboration, and self-efficacy can pay dividends in

all areas of instruction. Thinking smarter, and not harder about how instruction is approached can

result in a more efficient, rigorous, and rewarding program.

Action  

The three attributes that make the study of science-related vocabulary challenging, and

therefore a problem of practice, include the conceptual underpinnings of these words, the

copious amount of terms to be learned by a science student, and the specific, explicit strategies

needed to access this vocabulary and its connected text. Of the three challenges, the use of

       

   

164  

improved instructional strategies is the most easily addressed by science teachers. As a change

agent, I will work to provide science teachers, and, in fact, other content area teachers, with the

strategies and approaches to vocabulary acquisition that are most applicable and practical for

their unique content areas. Emphasis will be placed on providing children with repeated exposure

to word parts that they will see in different contexts and contents with explicit explanations of

their concepts and constructs so that they might make useful associations and connections that

will help them in all subject areas, and, indeed in their lives beyond school.

       

   

165  

References

 Aaron, P.G., Joshi, R.M., and Quatroche, D. (2008). Becoming a Professional Reading Teacher.

Baltimore, MD: Paul H. Brookes Publishing.

Achieve, Inc. (February 2005). Rising to the Challenge: Are High School Graduates Prepared for College and Work? A Study of Recent High School Graduates, College Instructors, and Employers. Washington, D.C. Prepared by Peter D. Hart Research Associates/Public opinion Strategies.

Achieve, Inc. (2013). Explore the NGSS. Next Generation Science Standards: For States, By

States. Retrieved January 10, 2013 from http://www.nextgenscience.org ACT, Inc. (2005). ACT Profile Report-National: Graduating Class 2012. Retreived from:

http://www.act.org/readiness/2012 Advameg, Inc. (2014). Massachusetts Bigger Cities (over 6000 residents). City-Data.com.

Retrieved from http://www.city-data.com/city/Massachusetts.html Anderson R. C. & Freebody, P. (1981). Vocabulary knowledge. In J. T. Guthrie (Ed.),

Comprehension and teaching: research reviews (77-117). Newark, DE: International Reading Association.

Anderson, L.W. & Krathwohl, D.R. (eds). (2001). A Taxonomy for Learning, Teaching, and

Assessing: A Revision of Bloom’s Taxonomy of Educational Objectives. Boston, MA: Allyn and Bacon.

Barton, P.E. (January 2000). What Jobs Require: Literacy, Education, and Training, 1940-2006.

Princeton, NJ: Policy Information Center, Education Testing Service. Baumann, J.F., Kame’enui, E.J., & Ash, G.E. (2003). Research on vocabulary instruction:

Voltaire redux. In J. Flood, D. Lapp, J.R. Squire, & J. M. Jenson (Eds.), Handbook of research on teaching the English Language Arts (2nd ed., 752-785). Mahwah, NJ: Erlbaum.

Beck, I.L., McKeown, M. G. & Kucan, L. (2002). Bringing words to life. New York: Guilford

Press. Beck, I.L., McKeown, M. G. & Omanson, R. C. (1987). The effects and uses of diverse

vocabulary instructional techniques. In M. G. McKeown & M. E. Curtis (Eds.), The nature of vocabulary acquisition (147-163). Hillsdale, NJ: Erlbaum.

       

   

166  

Biancarosa, C., & Snow, C. E. (2004). Reading Next—A vision for action and research in middle and high school literacy: A report to Carnegie Corporation of New York. Washington, DC: Alliance for Excellent Education.

Biancarosa, C., & Snow, C. E. (2006). Reading Next—A vision for action and research in middle

and high school literacy: A report to Carnegie Corporation of New York (2nd ed.).Washington, DC: Alliance for Excellent Education.

Bloodgood, J.W. & Pacifici, L.C. (2004). “Bringing word study to intermediate classrooms.” The

Reading Teacher Vol. 58, No. 3 International Reading Association (250–263) doi:10.1598/RT.58.3.3

Bloom, B.S. & Krathwohl, D.R. (1956). Taxonomy of Educational Objectives: the Classification

of Educational Goals by a committee of college and university examiners. Handbook I: Cognitive Domain. New York, NY: Longmans, Green.

Carnevale, A. P. (2001). Help wanted…College required. Washington, DC: Educational Testing

Service, Office for Public Leadership. Carrier, K. (2005). Supporting science learning through science literacy objectives for English

language learners. Science Activities, 42(2), 5-11. doi:10.3200/SATS.42.2.5-11. Chall, J.S. & Jacobs, V. A. (2003). Poor children’s fourth-grade slump. American Educator,

Spring, 2003. American Federation of Teachers. Cheung, T. (May 2013). Relationships and convergences among the mathematics, science, and

ELA practices. Refined version of diagram created by the Understanding Language Initiative for ELP Standards. Palo Alto, CA: Stanford University.

Creswell, J.W. (2009). Research design: Qualitative, quantitative, and mixed method approaches

(3rd ed.). Thousand Oaks, CA: Sage Publications, Inc. Cruz, M.B. (2012). Student and Teacher Perceptions of a Mobile-Based Biology Vocabulary

Study Tool for English Language Learners. (Unpublished Doctoral Dissertation). George Mason University, Fairfax, Virginia.

D’Arcangelo, M. (2002). The challenge of content-area reading: A conversation with Donna

Ogle. Educational Leadership, 60(15), 12-15. Frayer, D., Fredrick, W.C., & Klausmeier, H.J. (1969). A Schema for Testing the Level of

Concept Mastery. Working Paper from the Wisconsin Research and Development Center for Cognitive Learning, The University of Wisconsin, No. 16.

Frayer, D., Ghatala, E.S., & Klausmeier, H.J. ((1972). Levels of Concept Mastery: Implications

for Instruction. Educational Technology, 12/12, 23-29.

       

   

167  

Graves, M. F. (1986). Vocabulary learning and instruction. In E. Z. Rothkopf (Ed.), Review of

Research in Education, 13, 49-89. Halliday, M.A.K. (1993). Some grammatical problems in scientific English. In Halliday, M.A.K

& Martin, J.R. (Eds.), Writing Science: Literacy and Discursive Power (69-85). Great Britain: University of Pittsburg Press.

Hiebert E. H. & Kamil, M. L. (2005). Teaching and learning vocabulary: Bringing research to

practice. Mahwah, NJ: Erlbaum. Hirsch, E.D. (2003). Reading comprehension requires knowledge—of words and the world:

Scientific insights into the fourth-grade slump and the nation’s stagnant comprehension scores. American Educator, Spring, 2003. American Federation of Teachers.

Institute of Education Sciences (IES). (August 2008). Improving Adolescent Literacy: Effective Classroom and Intervention Practices. Kamil, M.K. et al (panel). U.S. Department of Education: NCEE 2008-4027.

Jacobs, R. L. (2011). Developing a Research Problem and Purpose Statement. In D.S. Rocco, T.

Hatcher, and Associates. The Handbook of Scholarly Writing and Publishing. San Francisco: Jossey-Bass.

Krueger, R.A. and Casey, M.A. (2009). Focus Groups: A Practical Guide for Applied Research.

4th edition. Thousand Oaks, California: Sage Publications. Kluckhohn, F. (1940). The Participant Observer Technique in Small Communities. American

Journal of Sociology. 46, 331-343. Lacey, M. (1998). Describing and displaying data: Scatterplots. Statistical Topics. Retrieved

January 12, 2014 from http://www.stat.yale.edu/Courses/1997-98/101/stat101.htm LeCompte, M. and Preissle, J. (1993). Ethnography and Qualitative Design in Educational

Research. London: Academic Press, Inc. Lee, O., & Fradd, S.H. (1996). Literacy skills in science learning among linguistically diverse

students. Science Education, 80(6), 651-671. doi:10.1002/(SICI)1098 237X(199611)80:6<651::AID-SCE2>3.0.CO;2-I.

Lemke, J.L. (1990). Talking science: Language, learning and values. Norwood, NJ: Ablex. MA DESE (Massachusetts Department of Elementary and Secondary Education) (March 2011).

Massachusetts Curriculum Framework for English Language Arts and Literacy Grades Pre-Kindergarten to12. Incorporating the Common Core State Standards for English

       

   

168  

Language Arts and Literacy in History/Social Studies, Science, and Technical Subjects. Retrieved October 5, 2013 from: http://www.doe.mass.edu/frameworks/ela/0311.pdf

MA DESE (Massachusetts Department of Elementary and Secondary Education) (February

2013). Characteristics of Standards-Based Teaching and Learning: Continuum of Practice. Accountability, Partnership, & Assistance. Retrieved January 10, 2014 from http://www.doe.mass.edu/apa/dart/walk/

MA DESE (Massachusetts Department of Elementary and Secondary Education) (January 2014).

Science, Technology, Engineering, and Mathematics (STEM). Retrieved January 11, 2014 from http://www.doe.mass.edu/STEM/review.html

MA DESE (Massachusetts Department of Elementary and Secondary Education) (2014).

Enrollment Data. School and District Profiles. Retrieved from http:// profiles.doe.mass.edu/profiles/student.aspx?orgcode=06650000&orgtypecode=5&

Maxwell, J. A. (2005). Qualitative research design: An interactive approach (2nd ed.).

Thousand Oaks, CA: Sage Publications, Inc. McKeown, M.G., Beck, I.L., Omanson, R.C., & Perfetti, C.A. (1983). The effects of long-term

vocabulary instruction on reading comprehension: A replication. Journal of Reading Behavior, 15, 3–18.

Merriam, S. B. (1998). Qualitative research and case study applications in education (2nd ed.) San Francisco, CA: Jossey-Bass Publishers.

Merriam, S. (2009). Qualitative research: A guide to design and implementation. San Francisco,

CA: Jossey-Bass Publications. Mezynski, K. (1983). Issues concerning the acquisition of knowledge: Effects of vocabulary

training on reading comprehension. Review of Educational Research, 53, 253-279. Moliken, Paul, Senior Ed. (2008) Growing Your Vocabulary: Learning from Greek and Latin

Roots. (A and B). Prestwick House, Inc. Nagy, W. E. (1988). Teaching vocabulary to improve reading instruction. Newark, DE.

International Reading Association. Nagy, W. E. (2005). Why vocabulary instruction needs to be long-term and comprehensive. In E.

H. Hiebert & M. L. Kamil (Eds.), Teaching and learning vocabulary: Bringing research to practice. Mahwah, NJ: Lawrence Erlbaum.

Nagy, W.E., Herman, P.A., and Anderson, R.C. (1985). Learning Words from Context. Reading

Research Quarterly. Vol. 20 Number 2. International Reading Association.

       

   

169  

National Academy of Sciences. (1996). National Science Education Standards. Washington, D.C.: National Academy Press.

National Institute of Child Health and Human Development (NICHHD). (2000). Report of the

National Reading Panel. Teaching children to read: an evidence-based assessment of the scientific research literature on reading and its implications for reading instruction: Reports of the subgroups (NIH Publication No. 00-4754). Washington, DC: U.S. Government Printing Office.

Partnership for Reading. (2001). Vocabulary: Instructional Guidelines and Classroom Examples.

Reading Rockets. Article 3473. Retrieved from: http://www.readingrockets.org/article/ 3473.

Patton, M. Q. (1990). Qualitative evaluation and research methods (2nd ed.), Newbury Park, CA:

Sage Publications, Inc. Ponterotto, J.G. (2005). Qualitative Research in Counseling Psychology: A Primer on Research

Paradigms and Philosophy of Science. Journal of Counseling Psychology, Vol. 52, No. 2, 126-136.

Rasinski, T., et al. (2008). Greek and Latin Roots: Keys to Building Vocabulary. Huntington

Beach, CA: Shell Education. Rubin, H.J and Rubin, I.S. (2012). Qualitative Interviewing: The Art of Hearing Data. Thousand

Oaks, California: Sage Publications, Inc. Rupley, W.H., Logan, J.W., & Nichols, W.D. (1998/1999). Vocabulary instruction in a balanced

reading program. The Reading Teacher, 52 (4). Saldana, Johnny. (2009). The Coding Manual for Qualitative Researchers. Thousand Oaks,

California: Sage Publications. Scarborough, H. (2001). Connecting early language and literacy to later reading (dis)abilities:

Evidence, theory, and practice. In S.B. Neuman & D.K. Dickinson (Eds.), Handbook of Early Literacy Research. NY: Guilford Press.

Siebert, D. & Draper, R. J. (2008). Why content-area literacy messages do not speak to

mathematics teachers: a critical content analysis, literacy research and instruction. 47:4. Provo, Utah: Brigham Young University. doi: 10.1080/193880070802300314

Sinatra, R. C., Berg, D. & Dunn, R. (1985). Semantic mapping improves reading comprehension

of reading disabled students. Teaching Exceptional Children, 17(4), 310-314. Stahl, S. A. (1983). Differential word knowledge and reading comprehension. Journal of

Reading Behavior, 15(4), 33-50.

       

   

170  

Stahl, S. A. (1991). Beyond the instrumentalist hypothesis: Some relationships between word

meanings and comprehension. In P. J. Schwanenflugel (ed.), The psychology of word meanings (157-186). Hillsdale, NJ: Erlbaum.

Stahl, S. A. & Fairbanks, M. M. (1986). The effects of vocabulary instruction: A model-based

meta-analysis. Review of Educational Research, 56, 72-110. Stanovich, K. E. (1986). Matthew effects in reading: Some consequences of individual

differences in the acquisition of literacy. Reading Research Quarterly, 21, 36–407. Thomas, D.R. (2006). A General Inductive Approach for Analyzing Qualitative Evaluation Data.

American Journal of Evaluation. Vol. 27, No. 2, 237-246. Tomlinson, C. A. (2010). Dynamic Curriculum: Differentiation Central. [Video file]. Retrieved

from http://www.differentiationcentral.com/videos.html Wandersee, J. H. (1985). Plants or animals—which do junior high school students prefer to

study? Journal of research in Science Teaching. Vol. 23, Issue 5. Weiss, R. S. (1994). Learning from strangers: The art and method of qualitative interview

studies. New York: The Free Press.

       

   

171  

Appendix  A  –  Grade  4  cumulative  assessment,  lessons  1-­‐5

       

   

172  

       

   

173  

Appendix B – Grade 5 cumulative assessment, lessons 1-5

       

   

174  

       

   

175  

       

   

176  

       

   

177  

Appendix  C  –  Nonsense  words  for  grades  four  and  five  assessments

GRADE 4 NONSENSE WORDS

ELA

Science

Both

Word parts

Aquaskates Exorain Moneymeter Auto Hydrocar Invertebracow Thermoshoes Aqua Astropants Millistars Symmetrishapes Hydr Lunafood Amphibicat Solarexplorer Astr Terrahouse Luna Terr Exo Vertebra Milli Ex: autopencil Amph Meter Thermo Sym Sol(ar)

GRADE 5 NONSENSE WORDS

ELA

Science

Both

Word parts

Florabook Biohouse Omnicolors Frater Trifeet Electrodog Respihelmet Matr Unistar Visiphone Chocolovore Flor(a) Matricake Photofood Subdirt Uni Duoflag Duo Tri Bio Electro Ex: fratball Vis Photo Omni Resp Vor Sub

       

   

178  

Appendix D – Multiple-choice assessment for grade four Name:________________________________ Date:___________ Teacher:__________________________

Science vocabulary DIRECTIONS: Choose the best answer for each question, and write the letter of your answer in the space provided.

1. Which of the following things will hydrate you? 1. _______ A. Cookies B. Bread C. Water D. Crackers E. Toast

2. The words millimeter, decimeter, and kilometer are used when:

A. Growing things B. Collecting things 2. _______ C. Measuring things D. Insulating things E. Heating things

3. Energy from the sun is called: 3. _______

A. Mechanical energy B. Electrical energy C. Wind energy D. Sound energy E. Solar energy

4. Which word probably means growing plants with water and no soil?

A. Photogenic B. Thermonuclear 4. _______ C. Astrophysicist D. Hydroponics E. Centrifuge

5. The words astrology and constellatory both have to do with:

A. Stars B. The sun 5. _______ C. The earth D. Fire E. The moon

6. This is: 6. _______

       

   

179  

A. an amphibian B. a mammal C. a fish D. an arthropod E. a worm

7. Which of the following would you NOT see if you were looking at terrain?

A. Grass B. Dirt 7. _______ C. Stars D. Sand E. Mud

8. Dehydrate means: 8. _______

A. To lose or remove water B. To do something every year C. To be easily heard D. To measure something E. To move water in a container

9. Choose the sport that is aquatic: 9. _______

A. Baseball B. Swimming C. Soccer D. Football E. Basketball

10. The word parts ecto and exo both refer to the: 10._______

A. Inside B. Upper C. Lower D. Outside E. Under

11. Solar and lunar refer to things that are found in the:

A. Ground 11._______ B. Water C. Dirt D. Sky E. Sand

12. You can NOT measure the diameter of: 12._______ A. a pizza B. a circle C. a house D. a coin E. a wheel

       

   

180  

13. An amphibious car can probably: 13._______ A. Drive on land only B. Drive in the water only C. Drive on both land and water D. Drive on rocks only E. Drive on dirt and sand

14. Which shape has symmetry? 14._______

A.

B.

C.

D.

E.

15. Which one has NO backbone? 15._______ A. Snake B. Human C. Dog D. Bird E. Fish

16. Which of the following is NOT an example of something thermal?

A. An ice cube B. The sun 16._______ C. A fire D. A heated oven E. A lit match

17. A vertebrate is: 17._______ A. An animal that has gills B. An animal that has no backbone C. An animal that has claws D. An animal that has a backbone E. An animal that has lungs

18. This instrument is used to measure temperature: 18._______

A. Thermometer B. Barometer

       

   

181  

C. Centimeter D. Metric ruler E. Balance scale

19. An astronomer studies: 19._______

A. Stars and planets B. Food and drink C. Plants and animals D. Rocks and minerals E. Liquids and solids

20. An arthropod has: 20._______

A. a vertebra B. an exoskeleton C. hollow bones D. an endoskeleton E. oxygenated blood

       

   

182  

Appendix E – Multiple-choice assessment for grade five Name:________________________________ Date:___________ Teacher:__________________________

Science vocabulary DIRECTIONS: Choose the best answer for each question, and write the letter of your answer in the space provided.

1. Which of the following things would a florist sell? F. Food and drinks 1. _______ G. Pens and pencils H. Dogs and cats I. Rocks and minerals J. Flowers and plants

2. Which of the following would NOT be eaten by an herbivore?

A. Apple B. Hamburger 2. _______ C. Carrot D. Lettuce E. Peanuts

3. A combination of electrical and magnetic energy is:

A. Electricity 3. _______ B. Chemical energy C. Magnetism D. Electromagnetic energy E. Nuclear energy

4. The work of biochemists and bioengineers probably has to do with:

A. Life or living things B. The solar system 4. _______ C. Family relationships D. Technology E. Electricity

5. Visible and visionary both refer to what someone can:

A. Hear B. Seen 5. _______ C. Say D. Feel E. Smell

6. The process in which plans take in light, carbon, dioxide, and water to produce sugar and oxygen

is: A. Photosynthesis 6. _______

       

   

183  

B. Chlorophyll C. Cell respiration D. Transpiration E. Adaptation

7. Which is an example of dual shapes?

A. 7. _______

B.

C.

D.

E.

8. Meat-eating animals usually have: A. Soft fur and tails 8. _______ B. Small teeth and tails C. Scales and fins D. Large teeth and claws E. Small eyes and no claws

9. Subsoil is:

A. The layer of soil above the topsoil 9. _______ B. The same thing as the earth’s crust C. The same thing as topsoil D. The layer of soil under the topsoil E. The layer of earth on top of the crust

10. Which is NOT a biotic factor?

A. Car 10._______ B. Plant C. Animal D. Tree E. Human being

11. The words unite, union, and unison refer to doing things: A. with a group of people B. with more than three people 11._______ C. with no help from anyone D. together as one E. as the leader

       

   

184  

12. These use what kind of energy? 12._______ A. Electrical energy B. Potential energy C. Chemical energy D. Physical energy E. Kinetic energy

13. Cell respiration requires:

A. Fire 13._______ B. Oxygen C. Prey D. Water E. Pollination

14. Humans are:

A. Decomposers 14._______ B. Herbivores C. Omnivores D. Producers E. Carnivores

15. Flora is the probably the goddess of:

A. Stars 15._______ B. Water C. Flowers D. Beaches E. Rocks

16. A collection of three works of art that are related to each other is:

A. a duet B. a duplicate 16._______ C. a trilogy D. a tricep E. a union

17. If someone is having trouble breathing, they probably have which kind of problem?

A. Circulatory 17._______ B. Respiratory C. Muscular D. Cellular E. Skeletal

18. The words tripetalous and trilayer both have to do with:

A. Living things B. Three parts 18._______ C. Mothers and fathers

       

   

185  

D. The solar system E. Dark colors

19. Which of the following probably does NOT have to do with light?

A. Photosynthesis B. Photon 19._______ C. Photograph D. Planetary E. Photocopy

20. Which of the following is NOT a word that has to do with family or family members?

A. Maternal 20._______ B. Paternity C. Thermal D. Matriarch E. Fraternal

       

   

186  

Appendix  F  –  Grade  Four  Student  Assessment  Questions  by  Type

Developed from (Frayer, D., Fredrick, W., Klausmeier, H., 1969) and (Frayer, D., Ghatala, E., Klausmeier, F. 1972)

Type of question

Prototype Question Vocabulary word/part

Discriminating Attributes

Given the name of an attribute value, the student can select the example of the attribute value.

Ø Which one has NO backbone?

F. Snake G. Human H. Dog I. Bird J. Fish

Invertebrate

Discriminating Attributes

Given an example of the attribute value, the student can select the name of the attribute value.

Ø Which word probably means growing plants with water and no soil?

F. Photogenic G. Thermonuclear H. Astrophysicist I. Hydroponics J. Centrifuge

Hydro

Classificatory Given the name of a concept, the student can select the example of the concept.

Ø Which shape has symmetry?

F.

G.

H.

I.

Symmetry

       

   

187  

J.

Ø Which of the following things will hydrate you?

K. Cookies L. Bread M. Water N. Crackers O. Toast

Hydrate

Classificatory Given an example of a concept, the student can select the name of the concept.

Ø This is: F. an amphibian G. a mammal H. a fish I. an arthropod J. a worm

Ø An arthropod has: F. a vertebra G. an exoskeleton H. hollow bones I. an endoskeleton J. oxygenated blood

Amphibian Exoskeleton

Formal Given the name of a concept, the student can select the names of the relevant attribute values of the concept.

Ø An amphibious car can probably:

A. Drive on land only B. Drive in the water

only C. Drive on both land

and water D. Drive on rocks only E. Drive on dirt and

sand

Ø The words millimeter, decimeter, and kilometer are used when:

F. Growing things G. Collecting things

Amphibious Millimeter, decimeter, kilometer

       

   

188  

H. Measuring things I. Insulating things J. Heating things

Ø An astronomer

studies: F. Stars and planets G. Food and drink H. Plants and animals I. Rocks and minerals J. Liquids and solids

Astronomer

Formal Given the name of a concept, the student can select the names of the irrelevant attribute values of the concept.

Ø Which of the following would you NOT see if you were looking at terrain?

F. Grass G. Dirt H. Stars I. Sand J. Mud

Ø You can NOT

measure the diameter of:

F. a pizza G. a circle H. a house I. a coin J. a wheel

Terrain Diameter

Formal Given the definition of a concept, the student can select the name for the concept.

Ø This instrument is used to measure temperature:

F. Thermometer G. Barometer H. Centimeter I. Metric ruler J. Balance scale

Ø Energy from the sun

is called: F. Mechanical energy G. Electrical energy H. Wind energy

Thermometer Solar

       

   

189  

I. Sound energy J. Solar energy

Formal Given the name of a

concept, the student can select the correct definition of the concept.

Ø Dehydrate means: F. To lose or remove

water G. To do something

every year H. To be easily heard I. To measure

something J. To move water in a

container

Ø A vertebrate is: F. An animal that has

gills G. An animal that has

no backbone H. An animal that has

claws I. An animal that has a

backbone J. An animal that has

lungs

(De)hydrate Vertebrate

Superordinate-Subordinate

Given the name of a concept and a concept coordinate to it, the student can select the name of a concept supraordinate to it.

Ø Solar and lunar refer to things that are found in the:

F. Ground G. Water H. Dirt I. Sky J. Sand

Solar/Lunar

Superordinate-Subordinate

Given the name of a concept, the student can select the name of a concept subordinate to it.

Ø Choose the sport that is aquatic:

F. Baseball G. Swimming H. Soccer I. Football J. Basketball

Ø Which of the

following is NOT an

Aqua(tic) Thermal

       

   

190  

example of something thermal?

F. An ice cube G. The sun H. A fire I. A heated oven J. A lit match

Principle Given the name of

two concepts, the student can select the principle which relates them.

Ø The words astrology and constellatory both have to do with:

A. Stars B. The sun C. The earth D. Fire E. The moon

Ø The word parts ecto

and exo both refer to the:

F. Inside G. Upper H. Lower I. Outside J. Under

Astrology/ Constellatory Ecto/exo

       

   

191  

Appendix  G  –  Grade  Five  Student  Assessment  Questions  by  Type Developed from (Frayer, D., Fredrick, W., Klausmeier, H., 1969) and (Frayer, D., Ghatala, E., Klausmeier, F. 1972)

Type of question

Prototype Question Vocabulary word/part

Discriminating Attributes

Given the name of an attribute value, the student can select the example of the attribute value.

Ø Meat-eating animals usually have:

K. Soft fur and tails L. Small teeth and tails M. Scales and fins N. Large teeth and claws O. Small eyes and no

claws

Carnivore

Discriminating Attributes

Given an example of the attribute value, the student can select the name of the attribute value.

Ø If someone is having trouble breathing, they probably have which kind of problem?

K. Circulatory L. Respiratory M. Muscular N. Cellular O. Skeletal

Respiratory

Classificatory Given the name of a concept, the student can select the example of the concept.

Ø Which is an example of dual shapes?

K.

L.

M.

N.

Dual

       

   

192  

O.

Ø Which of the following things would a florist sell?

P. Food and drinks Q. Pens and pencils R. Dogs and cats S. Rocks and minerals T. Flowers and plants

Florist

Classificatory Given an example of a concept, the student can select the name of the concept.

Ø These use what kind of energy?

K. Electrical energy L. Potential energy M. Chemical energy N. Physical energy O. Kinetic energy

Ø Humans are: K. Decomposers L. Herbivores M. Omnivores N. Producers O. Carnivores

Electrical Omnivores

Formal Given the name of a concept, the student can select the names of the relevant attribute values of the concept.

Ø Flora is the probably the goddess of:

F. Stars G. Water H. Flowers I. Beaches J. Rocks

Ø The words unite,

union, and unison refer to doing things:

K. with a group of

Flora Unite, union, unison

       

   

193  

people L. with more than three

people M. with no help from

anyone N. together as one O. as the leader

Ø Cell respiration

requires: K. Fire L. Oxygen M. Prey N. Water O. Pollination

Respiration

Formal Given the name of a concept, the student can select the names of the irrelevant attribute values of the concept.

Ø Which of the following would NOT be eaten by an herbivore?

K. Apple L. Hamburger M. Carrot N. Lettuce O. Peanuts

Ø Which is NOT a

biotic factor? K. Car L. Plant M. Animal N. Tree O. Human being

Herbivore Biotic

Formal Given the definition of a concept, the student can select the name for the concept.

Ø The process in which plans take in light, carbon, dioxide, and water to produce sugar and oxygen is:

K. Photosynthesis L. Chlorophyll M. Cell respiration N. Transpiration O. Adaptation

Photosynthesis

       

   

194  

Ø A collection of three works of art that are related to each other is:

K. a duet L. a duplicate M. a trilogy N. a tricep O. a union

Trilogy

Formal Given the name of a concept, the student can select the correct definition of the concept.

Ø Subsoil is: K. The layer of soil

above the topsoil L. The same thing as

the earth’s crust M. The same thing as

topsoil N. The layer of soil

under the topsoil O. The layer of earth on

top of the crust

Ø A combination of electrical and magnetic energy is:

K. Electricity L. Chemical energy M. Magnetism N. Electromagnetic

energy O. Nuclear energy

Subsoil Electromagnetic

Superordinate-Subordinate

Given the name of a concept and a concept coordinate to it, the student can select the name of a concept supraordinate to it.

Ø Visible and visionary both refer to what someone can:

K. Hear L. Seen M. Say N. Feel O. Smell

Visible, visionary

Superordinate-Subordinate

Given the name of a concept, the student can select the name of a

Ø Which of the following probably does NOT have to do with light?

Photosynthesis, photon, photograph, photocopy

       

   

195  

concept subordinate to it.

K. Photosynthesis L. Photon M. Photograph N. Planetary O. Photocopy

Ø Which of the following is NOT a word that has to do with family or family members?

K. Maternal L. Paternity M. Thermal N. Matriarch O. Fraternal

Maternal, Paternity, Matriarch, Fraternal

Principle Given the name of two concepts, the student can select the principle which relates them.

Ø The words tripetalous and trilayer both have to do with:

F. Living things G. Three parts H. Mothers and fathers I. The solar system J. Dark colors

Ø The work of

biochemists and bioengineers probably has to do with:

K. Life or living things L. The solar system M. Family relationships N. Technology O. Electricity

Tripetalous and Trilayer Biochemist/ bioengineer

       

   

196  

Appendix  H  –  Nonsense  word  assessment  for  grade  four Name:_____________________________ Date:__________ Teacher:_____________________

That’s Nonsense! DIRECTIONS: Each word below is a nonsense word (a make-believe word). See if you can figure out what the word might mean by breaking it down into its different parts. Draw a picture and then describe your picture.

Nonsense word

My picture

It looks like this because…

EXAMPLE:

autopencil

A pencil that writes by itself

aquaskates

exorain

 

       

   

197  

 

Nonsense word

My picture

It looks like this because…

moneymeter

hydrocar

invertebracow

thermoshoes

   

       

   

198  

Nonsense word

My picture

It looks like this because…

astropants

millistars

symmetrishapes

lunafood

   

       

   

199  

Nonsense word

My picture

It looks like this because…

amphibicat

solarexplorer

terrahouse

       

   

200  

Appendix I – Nonsense word assessment for grade five Name:_____________________________ Date:_________ Teacher:______________________

That’s Nonsense! DIRECTIONS: Each word below is a nonsense word (a make-believe word). See if you can figure out what the word might mean by breaking it down into its different parts. Draw a picture and then describe your picture.

Nonsense word

My picture

It looks like this because…

EXAMPLE:

fratball

Me playing ball with my brother

florabook

biohouse

   

       

   

201  

Nonsense word

My picture

It looks like this because…

omnicolors

trifeet

electrodog

respihelmet

   

       

   

202  

Nonsense word

My picture

It looks like this because…

unistar

visiphone

chocolovore

matricake

   

       

   

203  

Nonsense word

My picture

It looks like this because…

photofood

subdirt

duoflag

       

   

204  

Appendix  J  –  Focus  Group  Questions Four Focus Groups: Grade 4 ELA teachers and special educators, Grade 5 ELA teachers and special educators, Grade 4 Science teachers and special educators, Grade 5 Science teachers and special educators.

Type

Questions

Opening Questions

How long have each of you been teaching? At which school(s) have you taught?

Introduction Questions

How long have you been teaching vocabulary using Greek & Latin word parts? How did you choose and teach vocabulary prior to this? What type of teacher collaboration (ELA and science) is currently present in sharing vocabulary?

Transition Questions

What were your first impressions of using Greek and Latin as part of vocabulary instruction? Has that impression changed? If so, how? What are some instructional strategies that you use? What is their level(s) of success?

Key Questions Have you perceived any changes in student performance in vocabulary acquisition? If so, please describe. Have you perceived any changes in student performance in other literacy skills (fluency, comprehension, etc.)? If so, please describe. Have you noticed any other changes (behavior, motivation, engagement, etc.)? Do you have any evidence of changes (quantitative data, anecdotes, observations, etc.)?

Ending Questions

Are there any changes you would like to make with the vocabulary instruction or content you currently use? If so, what? Is there anything I should have/could have asked that you would like to add?

Follow-up Questions

Based on answers to the above questions

Question types based on Focus Group model by Krueger and Casey (2009).

       

   

205  

Appendix  K  –  Individual  Interview  Questions

Two Individual Interviews: ELA Curriculum Leader, Science Curriculum Leader

Type

Questions (ELA)

Tour Questions Please describe how a typical vocabulary unit is organized from start to finish. Please describe your experience(s) incorporating Greek and Latin into your class(es)?

Main Questions What are problems you have encountered in your vocabulary units? What are successes you have encountered in your vocabulary units? How, if at all, do students apply their understanding of Greek and Latin word parts to acquire vocabulary and/or understand new concepts in ELA? How, if at all, do students hypothesize the meaning of unknown vocabulary utilizing their understanding of Greek and Latin word parts?

Follow up Questions

Based on answers to the above questions

Probes Based on answers to the above questions and the direction of the conversation

Type

Questions (Science)

Tour Questions Please describe how you incorporate vocabulary instruction into a science unit from start to finish. Please describe your experience(s) incorporating Greek and Latin into your class(es)?

Main Questions What are problems you have encountered in your vocabulary instruction? What are successes you have encountered in your vocabulary instruction? How, if at all, do students apply their understanding of Greek and Latin word parts to acquire vocabulary and/or understand new concepts in science? How, if at all, do students hypothesize the meaning of unknown science vocabulary utilizing their understanding of Greek and Latin word parts?

Follow up Questions

Based on answers to the above questions

Probes Based on answers to the above questions and the direction of the conversation Question types based on Responsive Interview model by Rubin and Rubin (2012).

       

   

206  

Appendix L – Answer keys for multiple-choice assessments (Grades four and five)

Grade 4 answer key: Grade 5 answer key:

1. C 1. E 2. C 2. B 3. E 3. D 4. D 4. A 5. A 5. B 6. A 6. A 7. C 7. B 8. A 8. D 9. B 9. D 10. D 10. A 11. D 11. D 12. C 12. A 13. C 13. B 14. A 14. C 15. A 15. C 16. A 16. C 17. D 17. B 18. A 18. B 19. A 19. D 20. B 20. C

       

   

207  

Appendix M – Answer keys for nonsense word assessments (First cycle coding for grades four and five)

Grade 4 answer key: 1. aquaskates – water 2. exorain – outside 3. moneymeter – measure/how much/counts 4. hydrocar – water/moisture 5. invertabracow – backbone/spine 6. thermoshoes – heat/temperature 7. astropants – star 8. millistars – thousands 9. symmetrishapes – similar/same 10. lunafood – moon 11. amphibicat – both (water and land) 12. solarexplorer – sun 13. terrahouse – earth/soil/land

Grade 5 answer key:

1. florabook – flower/plant 2. biohouse – life 3. omnicolors – all/many/multiple 4. trifeet – three 5. electrodog – electric(ity) 6. respihelmet – breathing/air 7. unistar – one/single 8. visiphone – see/sight/look 9. chocolovore – one that eats/feeds 10. matricake – mother 11. photofood – light 12. subdirt – under/below 13. duoflag – two/double

       

   

208  

Appendix N – Nonsense word assessment cycle one coding for all grade four classes

GRADE 4 – Class 4A 4A Code/Totals

? √ X O Total ques

Ques 1 ELA 2 35 0 0 37 Ques 2 Science 11 3 21 2 37

Ques 3 Both 14 14 9 0 37 Ques 4 ELA 2 19 14 2 37

Ques 5 Science 2 27 8 0 37 Ques 6 Both 4 23 8 2 37

Ques 7 ELA 22 3 10 2 37 Ques 8 Science 4 2 28 3 37

Ques 9 Both 25 6 4 2 37 Ques 10 ELA 10 8 15 4 37

Ques 11 Science 20 7 9 1 37 Ques 12 Both 10 10 15 2 37

Ques 13 ELA 1 4 30 2 37

TOTALS 127 161 171 22 459

0  

5  

10  

15  

20  

25  

30  

35  

40  

Ques  1    ELA  

Ques  2    Science  

Ques  3    Both  

Ques  4    ELA  

Ques  5    Science  

Ques  6    Both  

Ques  7    ELA  

Ques  8    Science  

Ques  9    Both  

Ques  10    ELA  

Ques  11    Science  

Ques  12    Both  

Ques  13    ELA  

?   √   X   O  

       

   

209  

GRADE 4 – Class 4B 4B Code/Totals

? √ X O Total ques

Ques 1 ELA 2 36 0 1 39

Ques 2 Science 6 7 23 3 39

Ques 3 Both 7 22 9 1 39

Ques 4 ELA 4 15 17 3 39

Ques 5 Science 1 27 8 3 39

Ques 6 Both 3 28 6 2 39

Ques 7 ELA 28 1 7 3 39

Ques 8 Science 16 1 18 4 39

Ques 9 Both 26 4 4 5 39

Ques 10 ELA 11 16 9 3 39

Ques 11 Science 18 11 7 3 39

Ques 12 Both 22 9 5 3 39

Ques 13 ELA 4 7 23 5 39

TOTALS 148 184 136 39 468

0  

5  

10  

15  

20  

25  

30  

35  

40  

Ques  1    ELA  

Ques  2    Science  

Ques  3    Both  

Ques  4    ELA  

Ques  5    Science  

Ques  6    Both  

Ques  7    ELA  

Ques  8    Science  

Ques  9    Both  

Ques  10    ELA  

Ques  11    Science  

Ques  12    Both  

Ques  13    ELA  

?   √   X   O  

       

   

210  

GRADE 4 – Class 4C

4C Code/Totals

? √ X O Total ques

Ques 1 ELA 2 35 3 0 40

Ques 2 Science 6 3 29 2 40

Ques 3 Both 14 17 9 0 40

Ques 4 ELA 4 8 27 1 40

Ques 5 Science 3 7 29 1 40

Ques 6 Both 4 29 7 0 40

Ques 7 ELA 31 1 8 0 40

Ques 8 Science 11 1 28 0 40

Ques 9 Both 29 4 7 0 40

Ques10 ELA 5 6 29 0 40

Ques11 Science 15 18 7 0 40

Ques12 Both 28 4 8 0 40

Ques13 ELA 3 6 31 0 40

TOTALS 155 139 222 4 516

0  

5  

10  

15  

20  

25  

30  

35  

40  

Ques  1    ELA  

Ques  2    Science  

Ques  3    Both  

Ques  4    ELA  

Ques  5    Science  

Ques  6    Both  

Ques  7    ELA  

Ques  8    Science  

Ques  9    Both  

Ques10  ELA  

Ques11  Science  

Ques12    Both  

Ques13  ELA  

?   √   X   O  

       

   

211  

GRADE 4 – Class 4D 4D Code/Totals

? √ X O Total ques

Ques 1 ELA 1 37 2 0 40

Ques 2 Science 4 1 22 13 40

Ques 3 Both 15 18 6 1 40

Ques 4 ELA 7 21 12 0 40

Ques 5 Science 1 21 12 6 40

Ques 6 Both 3 31 6 0 40

Ques 7 ELA 28 6 5 1 40

Ques 8 Science 9 0 26 5 40

Ques 9 Both 27 2 7 4 40

Ques 10 ELA 13 10 13 4 40

Ques 11 Science 21 16 3 0 40

Ques 12 Both 27 12 1 0 40

Ques 13 ELA 6 3 26 5 40

TOTALS 162 178 141 39 481

0  

5  

10  

15  

20  

25  

30  

35  

40  

Ques  1    ELA  

Ques  2    Science  

Ques  3    Both  

Ques  4    ELA  

Ques  5    Science  

Ques  6    Both  

Ques  7    ELA  

Ques  8    Science  

Ques  9    Both  

Ques  10    ELA  

Ques  11    Science  

Ques  12    Both  

Ques  13    ELA  

?   √   X   O  

       

   

212  

GRADE 4 – Class 4E 4E Code/Totals

? √ X O Total ques

Ques 1 ELA 4 29 5 0 38

Ques 2 Science 4 6 27 1 38

Ques 3 Both 8 21 8 1 38

Ques 4 ELA 1 19 17 1 38

Ques 5 Science 5 10 22 1 38

Ques 6 Both 6 25 6 1 38

Ques 7 ELA 25 2 10 1 38

Ques 8 Science 8 1 28 1 38

Ques 9 Both 29 0 8 1 38

Ques 10 ELA 9 8 20 1 38

Ques 11 Science 24 11 2 1 38

Ques 12 Both 27 6 4 1 38

Ques 13 ELA 2 9 26 1 38

TOTALS 152 147 183 12 482

0  

5  

10  

15  

20  

25  

30  

35  

Ques  1    ELA  

Ques  2    Science  

Ques  3    Both  

Ques  4    ELA  

Ques  5    Science  

Ques  6    Both  

Ques  7    ELA  

Ques  8    Science  

Ques  9    Both  

Ques  10    ELA  

Ques  11    Science  

Ques  12    Both  

Ques  13    ELA  

?   √   X   O  

       

   

213  

Appendix O – Nonsense word assessment cycle one coding for all grade five classes

GRADE 5 – Class 5A 5A Code/Totals

? √ X O Total ques Ques 1 ELA 2 27 8 0 37 Ques 2 Science 13 17 7 0 37 Ques 3 Both 12 17 5 3 37 Ques 4 ELA 0 28 8 1 37 Ques 5 Science 7 26 4 0 37 Ques 6 Both 4 12 17 4 37 Ques 7 ELA 3 20 11 3 37 Ques 8 Science 10 16 10 1 37 Ques 9 Both 10 12 10 5 37 Ques 10 ELA 5 4 24 4 37 Ques 11 Science 3 9 25 0 37 Ques 12 Both 6 15 16 0 37 Ques 13 ELA 2 28 6 1 37 TOTALS 77 231 151 22 459

0  

5  

10  

15  

20  

25  

30  

Ques  1    ELA  

Ques  2    Science  

Ques  3    Both  

Ques  4    ELA  

Ques  5    Science  

Ques  6    Both  

Ques  7    ELA  

Ques  8    Science  

Ques  9    Both  

Ques  10    ELA  

Ques  11    Science  

Ques  12    Both  

Ques  13    ELA  

?   √   X   O  

       

   

214  

GRADE 5 – Class 5B 5B Code/Totals

? √ X O Total ques

Ques 1 ELA 2 30 1 1 34

Ques 2 Science 9 12 7 6 34

Ques 3 Both 16 4 3 11 34

Ques 4 ELA 0 31 0 3 34

Ques 5 Science 8 24 1 1 34

Ques 6 Both 2 8 9 15 34

Ques 7 ELA 3 25 3 3 34

Ques 8 Science 7 11 5 11 34

Ques 9 Both 2 10 3 19 34

Ques 10 ELA 2 5 7 20 34

Ques 11 Science 9 9 11 5 34

Ques 12 Both 2 22 3 7 34

Ques 13 ELA 2 29 1 2 34

TOTALS 64 220 54 104 338

0  

5  

10  

15  

20  

25  

30  

35  

Ques  1    ELA  

Ques  2    Science  

Ques  3    Both  

Ques  4    ELA  

Ques  5    Science  

Ques  6    Both  

Ques  7    ELA  

Ques  8    Science  

Ques  9    Both  

Ques  10    ELA  

Ques  11    Science  

Ques  12    Both  

Ques  13    ELA  

?   √   X   O  

       

   

215  

GRADE 5 – Class 5C 5C Code/Totals

? √ X O Total ques

Ques 1 ELA 0 42 1 0 43

Ques 2 Science 11 16 15 1 43

Ques 3 Both 16 15 11 1 43

Ques 4 ELA 0 42 1 0 43

Ques 5 Science 14 28 0 1 43

Ques 6 Both 1 10 30 2 43

Ques 7 ELA 5 32 6 0 43

Ques 8 Science 25 14 3 1 43

Ques 9 Both 15 15 12 1 43

Ques 10 ELA 3 7 31 2 43

Ques 11 Science 7 6 29 1 43

Ques 12 Both 10 18 13 2 43

Ques 13 ELA 6 34 2 1 43

TOTALS 113 279 154 13 546

0  

5  

10  

15  

20  

25  

30  

35  

40  

45  

Ques  1    ELA  

Ques  2    Science  

Ques  3    Both  

Ques  4    ELA  

Ques  5    Science  

Ques  6    Both  

Ques  7    ELA  

Ques  8    Science  

Ques  9    Both  

Ques  10    ELA  

Ques  11    Science  

Ques  12    Both  

Ques  13    ELA  

?   √   X   O  

       

   

216  

GRADE 5 – Class 5D 5D Code/Totals

? √ X O Total ques

Ques 1 ELA 1 37 3 0 41

Ques 2 Science 19 10 12 0 41

Ques 3 Both 19 9 11 2 41

Ques 4 ELA 0 38 2 1 41

Ques 5 Science 15 23 2 1 41

Ques 6 Both 7 14 17 3 41

Ques 7 ELA 5 25 10 1 41

Ques 8 Science 16 13 11 1 41

Ques 9 Both 11 16 11 3 41

Ques 10 ELA 7 8 25 1 41

Ques 11 Science 10 1 28 2 41

Ques 12 Both 11 23 6 1 41

Ques 13 ELA 5 34 0 2 41

TOTALS 126 251 138 18 515

 

0  

5  

10  

15  

20  

25  

30  

35  

40  

Ques  1    ELA  

Ques  2    Science  

Ques  3    Both  

Ques  4    ELA  

Ques  5    Science  

Ques  6    Both  

Ques  7    ELA  

Ques  8    Science  

Ques  9    Both  

Ques  10    ELA  

Ques  11    Science  

Ques  12    Both  

Ques  13    ELA  

?   √   X   O  

       

   

217  

GRADE 5 – Class 5E 5E Code/Totals

? √ X O Total ques

Ques 1 ELA 0 32 6 0 38

Ques 2 Science 20 11 7 0 38

Ques 3 Both 18 14 5 1 38

Ques 4 ELA 2 36 0 0 38

Ques 5 Science 10 27 1 0 38

Ques 6 Both 2 9 25 2 38

Ques 7 ELA 8 21 9 0 38

Ques 8 Science 19 15 4 0 38

Ques 9 Both 9 19 10 0 38

Ques 10 ELA 2 10 26 0 38

Ques 11 Science 8 6 24 0 38

Ques 12 Both 7 19 12 0 38

Ques 13 ELA 1 36 1 0 38

TOTALS 106 255 130 3 491

0  

5  

10  

15  

20  

25  

30  

35  

40  

Ques  1    ELA  

Ques  2    Science  

Ques  3    Both  

Ques  4    ELA  

Ques  5    Science  

Ques  6    Both  

Ques  7    ELA  

Ques  8    Science  

Ques  9    Both  

Ques  10    ELA  

Ques  11    Science  

Ques  12    Both  

Ques  13    ELA  

?   √   X   O  

       

   

218  

Appendix P – Updated nonsense word assessment after cycle TWO coding for all grade four classes

 GRADE 4 – Class 4A

4A Code/Totals √ X O Total ques Ques 1 ELA 36 1 0 37 Ques 2 Science 5 30 2 37 Ques 3 Both 18 19 0 37 Ques 4 ELA 20 15 2 37 Ques 5 Science 28 9 0 37 Ques 6 Both 23 12 2 37 Ques 7 ELA 5 30 2 37 Ques 8 Science 2 32 3 37 Ques 9 Both 30 5 2 37 Ques 10 ELA 13 20 4 37 Ques 11 Science 18 18 1 37 Ques 12 Both 15 20 2 37 Ques 13 ELA 4 31 2 37 TOTALS 217 242 22 481

0  

5  

10  

15  

20  

25  

30  

35  

40  

Ques  1    ELA  

Ques  2    Science  

Ques  3    Both  

Ques  4    ELA  

Ques  5    Science  

Ques  6    Both  

Ques  7    ELA  

Ques  8    Science  

Ques  9    Both  

Ques  10    ELA  

Ques  11    Science  

Ques  12    Both  

Ques  13    ELA  

√   X   O  

       

   

219  

GRADE 4 – Class 4B 4B Code/Totals

√ X O Total ques

Ques 1 ELA 38 0 1 39

Ques 2 Science 11 25 3 39

Ques 3 Both 23 15 1 39

Ques 4 ELA 16 20 3 39

Ques 5 Science 27 9 3 39

Ques 6 Both 28 9 2 39

Ques 7 ELA 3 33 3 39

Ques 8 Science 2 33 4 39

Ques 9 Both 30 4 5 39

Ques 10 ELA 17 19 3 39

Ques 11 Science 26 10 3 39

Ques 12 Both 19 17 3 39

Ques 13 ELA 11 23 5 39

TOTALS 251 217 39 507

0  

5  

10  

15  

20  

25  

30  

35  

40  

Ques  1    ELA  

Ques  2    Science  

Ques  3    Both  

Ques  4    ELA  

Ques  5    Science  

Ques  6    Both  

Ques  7    ELA  

Ques  8    Science  

Ques  9    Both  

Ques  10    ELA  

Ques  11    Science  

Ques  12    Both  

Ques  13    ELA  

√   X   O  

       

   

220  

GRADE 4 – Class 4C 4C Code/Totals

√ X O Total ques

Ques 1 ELA 36 4 0 40

Ques 2 Science 6 32 2 40

Ques 3 Both 18 22 0 40

Ques 4 ELA 9 30 1 40

Ques 5 Science 10 29 1 40

Ques 6 Both 32 8 0 40

Ques 7 ELA 8 32 0 40

Ques 8 Science 1 39 0 40

Ques 9 Both 26 14 0 40

Ques 10 ELA 8 32 0 40

Ques 11 Science 33 7 0 40

Ques 12 Both 10 30 0 40

Ques 13 ELA 7 33 0 40

TOTALS 204 312 4 520

0  

5  

10  

15  

20  

25  

30  

35  

40  

45  

Ques  1    ELA  

Ques  2    Science  

Ques  3    Both  

Ques  4    ELA  

Ques  5    Science  

Ques  6    Both  

Ques  7    ELA  

Ques  8    Science  

Ques  9    Both  

Ques  10    ELA  

Ques  11    Science  

Ques  12    Both  

Ques  13    ELA  

√   X   O  

       

   

221  

GRADE 4 – Class 4D 4D Code/Totals

√ X O Total ques

Ques 1 ELA 37 3 0 40

Ques 2 Science 3 24 13 40

Ques 3 Both 20 19 1 40

Ques 4 ELA 26 14 0 40

Ques 5 Science 21 13 6 40

Ques 6 Both 31 9 0 40

Ques 7 ELA 14 25 1 40

Ques 8 Science 0 35 5 40

Ques 9 Both 23 13 4 40

Ques 10 ELA 11 25 4 40

Ques 11 Science 30 10 0 40

Ques 12 Both 25 15 0 40

Ques 13 ELA 8 27 5 40

TOTALS 249 232 39 520

0  

5  

10  

15  

20  

25  

30  

35  

40  

Ques  1    ELA  

Ques  2    Science  

Ques  3    Both  

Ques  4    ELA  

Ques  5    Science  

Ques  6    Both  

Ques  7    ELA  

Ques  8    Science  

Ques  9    Both  

Ques  10    ELA  

Ques  11    Science  

Ques  12    Both  

Ques  13    ELA  

√   X   O  

       

   

222  

GRADE 4 – Class 4E 4E Code/Totals

√ X O Total ques

Ques 1 ELA 33 5 0 38

Ques 2 Science 10 27 1 38

Ques 3 Both 21 16 1 38

Ques 4 ELA 19 18 1 38

Ques 5 Science 13 24 1 38

Ques 6 Both 26 11 1 38

Ques 7 ELA 9 28 1 38

Ques 8 Science 1 36 1 38

Ques 9 Both 23 14 1 38

Ques 10 ELA 9 28 1 38

Ques 11 Science 29 8 1 38

Ques 12 Both 11 26 1 38

Ques 13 ELA 10 27 1 38

TOTALS 214 268 12 494

0  

5  

10  

15  

20  

25  

30  

35  

40  

Ques  1    ELA  

Ques  2    Science  

Ques  3    Both  

Ques  4    ELA  

Ques  5    Science  

Ques  6    Both  

Ques  7    ELA  

Ques  8    Science  

Ques  9    Both  

Ques  10    ELA  

Ques  11    Science  

Ques  12    Both  

Ques  13    ELA  

√   X   O  

       

   

223  

Appendix Q – Updated nonsense word assessment after cycle TWO coding for all grade five classes

GRADE 5 – Class 5A 5A Code/Totals

√ X O Total ques

Ques 1 ELA 27 10 0 37

Ques 2 Science 23 14 0 37

Ques 3 Both 27 7 3 37

Ques 4 ELA 28 8 1 37

Ques 5 Science 29 8 0 37

Ques 6 Both 14 19 4 37

Ques 7 ELA 22 12 3 37

Ques 8 Science 19 17 1 37

Ques 9 Both 14 18 5 37

Ques 10 ELA 5 28 4 37

Ques 11 Science 9 28 0 37

Ques 12 Both 19 18 0 37

Ques 13 ELA 30 6 1 37

TOTALS 266 193 22 481

0  

5  

10  

15  

20  

25  

30  

35  

Ques  1    ELA  

Ques  2    Science  

Ques  3    Both  

Ques  4    ELA  

Ques  5    Science  

Ques  6    Both  

Ques  7    ELA  

Ques  8    Science  

Ques  9    Both  

Ques  10    ELA  

Ques  11    

Science  

Ques  12    Both  

Ques  13    ELA  

√   X   O  

       

   

224  

GRADE 5 – Class 5B 5B Code/Totals

√ X O Total ques

Ques 1 ELA 32 1 1 34

Ques 2 Science 15 13 6 34

Ques 3 Both 11 12 11 34

Ques 4 ELA 31 0 3 34

Ques 5 Science 30 3 1 34

Ques 6 Both 9 10 15 34

Ques 7 ELA 28 3 3 34

Ques 8 Science 15 8 11 34

Ques 9 Both 10 5 19 34

Ques 10 ELA 6 8 20 34

Ques 11 Science 17 12 5 34

Ques 12 Both 24 3 7 34

Ques 13 ELA 31 1 2 34

TOTALS 259 79 104 442

0  

5  

10  

15  

20  

25  

30  

35  

Ques  1    ELA  

Ques  2    Science  

Ques  3    Both  

Ques  4    ELA  

Ques  5    Science  

Ques  6    Both  

Ques  7    ELA  

Ques  8    Science  

Ques  9    Both  

Ques  10    ELA  

Ques  11    Science  

Ques  12    Both  

Ques  13    ELA  

√   X   O  

       

   

225  

GRADE 5 – Class 5C 5C Code/Totals

√ X O Total ques

Ques 1 ELA 42 1 0 43

Ques 2 Science 24 18 1 43

Ques 3 Both 26 16 1 43

Ques 4 ELA 42 1 0 43

Ques 5 Science 37 5 1 43

Ques 6 Both 10 31 2 43

Ques 7 ELA 33 10 0 43

Ques 8 Science 22 20 1 43

Ques 9 Both 18 24 1 43

Ques 10 ELA 9 32 2 43

Ques 11 Science 9 33 1 43

Ques 12 Both 26 15 2 43

Ques 13 ELA 40 2 1 43

TOTALS 338 208 13 559

0  

5  

10  

15  

20  

25  

30  

35  

40  

45  

Ques  1    ELA  

Ques  2    Science  

Ques  3    Both  

Ques  4    ELA  

Ques  5    Science  

Ques  6    Both  

Ques  7    ELA  

Ques  8    Science  

Ques  9    Both  

Ques  10    ELA  

Ques  11    Science  

Ques  12    Both  

Ques  13    ELA  

√   X   O  

       

   

226  

GRADE 5 – Class 5D 5D Code/Totals

√ X O Total ques

Ques 1 ELA 37 4 0 41

Ques 2 Science 21 20 0 41

Ques 3 Both 24 15 2 41

Ques 4 ELA 38 2 1 41

Ques 5 Science 32 8 1 41

Ques 6 Both 17 21 3 41

Ques 7 ELA 30 10 1 41

Ques 8 Science 20 20 1 41

Ques 9 Both 18 20 3 41

Ques 10 ELA 14 26 1 41

Ques 11 Science 4 35 2 41

Ques 12 Both 34 6 1 41

Ques 13 ELA 37 2 2 41

TOTALS 326 189 18 533

0  

5  

10  

15  

20  

25  

30  

35  

40  

Ques  1    ELA  

Ques  2    Science  

Ques  3    Both  

Ques  4    ELA  

Ques  5    Science  

Ques  6    Both  

Ques  7    ELA  

Ques  8    Science  

Ques  9    Both  

Ques  10    ELA  

Ques  11    Science  

Ques  12    Both  

Ques  13    ELA  

√   X   O  

       

   

227  

GRADE 5 – Class 5E 5E Code/Totals

√ X O Total ques

Ques 1 ELA 32 6 0 38

Ques 2 Science 23 15 0 38

Ques 3 Both 28 9 1 38

Ques 4 ELA 38 0 0 38

Ques 5 Science 33 5 0 38

Ques 6 Both 9 27 2 38

Ques 7 ELA 28 10 0 38

Ques 8 Science 24 14 0 38

Ques 9 Both 20 18 0 38

Ques 10 ELA 11 27 0 38

Ques 11 Science 10 28 0 38

Ques 12 Both 26 12 0 38

Ques 13 ELA 37 1 0 38

TOTALS 319 172 3 494

0  

5  

10  

15  

20  

25  

30  

35  

40  

Ques  1    ELA  

Ques  2    Science  

Ques  3    Both  

Ques  4    ELA  

Ques  5    Science  

Ques  6    Both  

Ques  7    ELA  

Ques  8    Science  

Ques  9    Both  

Ques  10    ELA  

Ques  11    Science  

Ques  12    Both  

Ques  13    ELA  

√   X   O