1,406
Views
0
CrossRef citations to date
0
Altmetric
Learning, Instruction, and Cognition

Impact of Lesson Study and Fractions Resources on Instruction and Student Learning

ORCID Icon, ORCID Icon, ORCID Icon, &

Abstract

Lesson study, especially when conducted with support from targeted resources, is often identified as a potentially effective means to support teacher learning. In this study, 80 school-based teams of educators, representing 80 classrooms of third- or fourth-grade students in the US, were randomly assigned to one of four conditions: lesson study supported by fractions resources, lesson study only, fractions resources only, or a comparison group. Fractions instruction was improved in the two conditions that received the fractions resources. Average student performance on a fractions test did not increase relative to the comparison group in any of the three intervention conditions. However, moderation analyses suggest that lesson study, both with and without fractions resources, may have improved student performance in one state, which represented half of the sample.

Fractions are a challenging topic for students (National Mathematics Advisory Panel, Citation2008; OECD, Citation2014). Fractions knowledge predicts proficiency in algebra, even when analyses adjust for other facets of mathematical knowledge, and fractions competence is often identified as a “gatekeeper” topic for secondary-level mathematics (Booth & Newton, Citation2012; Siegler et al., Citation2013). Only 20% of U.S. elementary teachers rate their own fractions knowledge as strong or very strong (Ward & Thomas, Citation2006), a finding that may explain why simply providing collegial work sessions for teachers to collaborate on improving fractions instruction is not sufficient to improve teachers’ mathematical knowledge of fractions (Gearhart et al., Citation1999; Saxe et al., Citation2001).

Despite billions of dollars spent every year on teacher professional development in the US and frequent claims that PD is an important lever by which to improve instruction and student learning, very few PD programs are rigorously evaluated for impact on classroom teaching and student learning (Fermanich, Citation2002; Garet et al., Citation2016; Hill et al., Citation2013; Odden et al., Citation2002; Schoen et al., Citation2018a; TNTP, Citation2015; U.S. Department of Education, Citation2014; Wei et al., Citation2010; Yoon et al., Citation2007). A 2014 report found that only two of 643 studies of mathematics PD found positive effects on student mathematics proficiency when the study design met rigorous scientific criteria (Gersten et al., Citation2014). Lesson study with support from a fractions resource kit (Lewis & Perry, Citation2017) was the subject of one of those two studies.

The study described here is a conceptual replication (Schmidt, Citation2009) of the Lewis and Perry (Citation2017) randomized trial. The intervention in the Lewis and Perry study involved mailing fractions-lesson study binders to teams of teachers who used them to conduct lesson study. The binders were designed to support teacher-led lesson study with the support of a set of resources to support teaching of fractions with an emphasis on linear representations of fractions concepts. Teacher and student-level results in the intervention condition were contrasted with those where teams of teachers conducted lesson study on a topic of their choosing or chose both the topic and a professional learning method other than lesson study. Lewis and Perry reported a positive impact of lesson study on both teachers’ and students’ fractions knowledge, but the study was not designed to separate the effects of the two main intervention components: the lesson study process and the fractions resources.

This conceptual replication study uses four treatment conditions—all of which were provided with some opportunity for teachers to learn about teaching fractions. This study was designed to enable evaluation of the independent and combined impacts of lesson study (a process for teacher learning) and content resources (fractions resources for teachers to study) on classroom instruction and student learning in fractions. This study also involved a larger sample, a slightly modified version of the measure of student outcomes, and a lead author and methodologist who represent a third-party (i.e., not the program developers or implementers) evaluation team. is provided to help orient the reader to some of the key similarities and differences between the current study and the Lewis and Perry study.

Table 1. Comparison of present study with Lewis and Perry (Citation2017) study.

Linear representations in fractions instruction

Linear representations of fractions, such as number line and distance measurement, may help students see fractions as numbers with magnitude, not just as objects in a set or pieces of a whole, allowing students to integrate knowledge of fractions and whole numbers into a unified number system (Lamon, Citation2005; Morris, Citation2000; Saxe et al., Citation2013; Siegler et al., Citation2010). Linear representations of fractions are emphasized in textbooks of some high-achieving Asian countries (Grow-Maienza & Beal, Citation2005; Ma, Citation1999; Tokyo Shoseki Co., Citation2015). Historically, U.S. textbooks have not emphasized linear representations (Lewis et al., Citation2011; Watanabe, Citation2007).

The Common Core State Standards for Mathematics (CCSSM) and related standards adopted by most U.S. states within the last decade emphasize the number line as a means to integrate fractions (as rational numbers) and whole numbers in elementary education. These standards also deemphasize use of discrete wholes (e.g., pizzas) and set models (e.g., 2 of 6 buttons) to represent fractions (NGACPB & CCSSO, Citation2010). Further, CCSSM emphasizes students’ foundational understanding of unit fractions (fractions with a numerator of one), the relationship of the unit fraction to non-unit fractions and to the referent unit (e.g., two-thirds is made up of two one-thirds; three one-thirds make one) and equipartitioning (dividing the referent unit into same-size unit fractions). In summary, the elementary-level CCSSM emphasize linear representations of fractions and the use of unit fractions as basic units to compose/decompose other numbers—changes that represent a departure from prior U.S. practice and a potential challenge for teachers to implement as intended (Common Core Standards Writing Team, Citation2013; Sidney et al., Citation2019). Several studies have evaluated the effects of interventions designed to incorporate linear representations of fractions into instruction using experimental designs and reported positive impact on student achievement on tests of fractions knowledge (Fuchs et al., Citation2017; Jayanthi et al., Citation2021; Lewis & Perry, Citation2017).

Lesson study

In lesson study—a teacher-driven approach to professional learning that is used widely in Japan—teachers conduct collaborative “Study-Plan-Teach-Reflect” cycles designed to improve classroom instruction (Lewis & Hurd, Citation2011; Takahashi & McDougal, Citation2014). The team of participating teachers may include all teachers at a given grade level, or, as in the current study, any team of interested volunteers. Teachers typically begin the cycle by studying the relevant content and teaching materials, laying out their immediate and long-term goals for students, and choosing one lesson within the planned unit to serve as a “research lesson” that will be collaboratively planned and then taught by one team member to that member’s students. Research lessons are so named, because teams enact their hypotheses about good instruction by designing a lesson, conducting the lesson, gathering data during implementation of the lesson, and discussing student responses to the lesson (Lewis et al., Citation2019). Revising and reteaching the lesson is an optional feature of lesson study, common in the U.S. but less so in Japan (Fujii, Citation2014).

Study of the academic content and teaching materials—kyouzai kenkyuu—is integral to lesson study in Japan (Takahashi et al, Citation2005; Yoshida & Jackson, Citation2011). During kyouzai kenkyuu, teachers study documents such as teacher’s manuals, content frameworks, and research reports to learn about both the subject matter and its teaching and learning (Doig et al., Citation2011; Takahashi et al., Citation2005). Japanese teacher’s manuals provide support for kyouzai kenkyuu by identifying and discussing the key mathematical ideas in each instructional unit, anticipating likely student solution strategies and linking them to the key mathematical ideas, and situating the unit within a multi-year trajectory of mathematical learning (Lee & Zusho, Citation2002; Lewis et al., Citation2011; Miyakawa, Citation2011; Takahashi, Citation2021). Lesson study engages teams of teachers in studying research-based instructional resources with a particular emphasis on planning their use and examining their impact on practice. Lewis and Perry (Citation2017) found that lesson study can increase U.S. teachers’ content knowledge for instruction and improve student learning.

Teaching materials found outside Japan do not always provide good support for kyouzai kenkyuu. For example, comparison of the treatment of quadrilateral area in the teacher’s manual for widely used Japanese and U.S. textbooks revealed that 28% of the statements in the Japanese teacher’s manual focused on student thinking, but only 1% of the statements in the U.S. manual focused on student thinking (Lewis et al., Citation2011). Similarly, 10% of the statements in the Japanese teacher’s manual focused on the rationale for instructional decisions (e.g., the choice not to show gridlines when asking students to consider how to find the area of a rectangle), whereas none of the statements in the U.S. teacher’s manual provided such rationale. This stark difference in curriculum-embedded professional-learning resources can limit the quality and coherence of various phases in the lesson-study cycle when U.S. teachers conduct lesson study.

Lesson study with fractions resources: Program theory of change

Enactment of new instructional approaches is difficult (Cohen, Citation2011), and fractions are a mathematically challenging topic for teachers (Ward & Thomas, Citation2006). In the intervention reported here and in the previous randomized trial (Lewis & Perry, Citation2017), the program theory of change focuses on two major elements hypothesized to improve fractions instruction and therefore student learning: (1) research-based resources on fractions that highlight the role of linear representations and (2) a lesson-study process for collaborating with colleagues to study the resources and improve fractions instruction. The research-based fractions resources were intended to expand the fractions knowledge available to teachers (especially related to linear representations and student thinking), and the lesson-study process was intended to support the challenging work of translating these resources into high-quality instruction. The theory of change suggests that the two elements together—resources and a collaborative process for studying them and using them to plan instruction—should be more effective at improving fractions teaching and student learning than either element in isolation.

Research questions

The study was designed to investigate the following research questions (RQs). We consider RQ1 and RQ2 to be confirmatory and RQ3 to be exploratory.

RQ1. What are the independent and combined impacts of lesson study and fractions resources on fractions instruction, as measured by the Instructional Quality Assessment (IQA)?

RQ2. What are the independent and combined impacts of lesson study and fractions resources on student learning in fractions, as measured by the Early Fractions Test (EFT)?

RQ3. To what extent are effects in fractions instruction and student achievement moderated by the following factors: (i) student (average) pretest score, (ii) grade level, (iii) geographic location, and (iv) average baseline fractions knowledge of the participating teacher team?

Method

The study reported here used a randomized controlled trial research design to investigate the impact of the lesson-study process, conducted with and without fractions mathematical resources. Eighty school-based teams of educators agreed to participate in the present study—about twice the number included in the study by Lewis & Perry (Citation2017)—but the teams were distributed across four experimental conditions (rather than three as in the prior trial). As described in detail below, at least one educator on each team was a third- or fourth-grade teacher responsible for fractions instruction during the study period. The teams were assigned to one of four conditions: lesson study supported by fractions resources (LS + FR); lesson study only (LS); fractions resources only (FR); and a comparison group (C).

Description of the intervention and counterfactual conditions

As in Lewis and Perry (Citation2017) study, the intervention reported here consisted of condition-specific binders mailed to participating sites in three intervention conditions and one control condition. The binder contents differed by experimental condition and were expected to catalyze different activities. The binders for the four conditions represent the four possible combinations of providing or not providing support in the form of (1) a professional learning process (lesson study) and (2) fractions content resources. Every participating teacher in all four conditions received their own binder. In all four conditions, the binders encouraged teachers to improve their fractions instruction and to log the time, materials, and processes they used to do so.

shows the binder contents by condition. The binder for the control condition provided the CCSSM fractions standards, suggested that teachers examine curriculum materials and research to improve their fractions instruction, and recommended some general resources (e.g., consulting a district math specialist, visiting the website of the National Council of Teachers of Mathematics). For the sites in the two conditions assigned to lesson study (i.e., LS + FR, LS), the binders also included step-by-step guidance on conducting a lesson study cycle, starting with developing norms and processes to work as a team, and then going through the four lesson-study phases: study the curriculum and content; plan a series of lessons; teach the lessons, where at least one lesson represents a research lesson that is observed by the whole team, and reflect on data from the lesson and learning during the whole cycle. For educators in the two conditions that received fractions resources (i.e., LS + FR, FR), the binder also included specific fractions resources (curriculum examples, video, and research summaries; see ) accompanied by discussion prompts (relabeled as “reflection” prompts in the FR condition, where participants were expected to study these materials individually). The selected fractions resources emphasized linear representation of fractions as a support for students in understanding fractions as numbers that can be placed on a number line (Saxe, Citation2007; Saxe et al., Citation2013; Watanabe, Citation2002, Citation2007; Wu, Citation2011). Each condition also received condition-specific restrictions, reproduced verbatim in .

Table 2. Binder contents by condition and condition-specific prompts and restrictions.

The binder sent to teachers in the LS + FR condition was similar to that used previously by Lewis & Perry (Citation2017). Two major additions were the inclusion of CCSSM excerpts on fractions and video of an additional series of lessons (at fourth grade). To accommodate these new materials while keeping binder length reasonable, several binder sections that were provided to sites in the previous study were eliminated, including five short investigations that teachers could conduct with their students.

Procedures

The target population for the study included teams of educators in U.S. public schools—regular or charter—with at least one educator responsible for teaching mathematics to third- or fourth-grade students.

Recruitment

Information about the opportunity to participate in the research study was advertised through two main channels. Every known school-district mathematics specialist and elementary-school principal in Florida was notified of the opportunity through an e-mail from the first author. The notice was also shared by the second author through several listservs focused on lesson study and through informal conversations at professional conferences. All the sites located outside Florida were recruited through the listservs and professional conferences. Potential participants were directed to an online survey, where they could access the consent form approved by the Institutional Review Board at Florida State University (FSU). When at least three educators (including at least one third- or fourth-grade teacher of mathematics) for a given school site had applied to participate and provided informed consent, research permission was requested from the relevant school district. Only one team per school site was allowed to participate.

Recruitment occurred in two cohorts. The initial cohort (Cohort 1) consisted of 68 sites that met the eligibility criteria for the 2016-2017 school year. Before randomization, we obtained positive parental consent and student assent for 1,211 students from these sites. The sites in this cohort were randomized in Fall 2016, after the student pretests were administered and the consent forms returned, and their associated data were collected during the 2016-2017 school year. To increase the number of sites in the study, a second cohort of 12 sites (with a total of 196 consenting students) that met the eligibility criteria (Cohort 2) was recruited to participate in the 2017-2018 school year. Apart from the year delay, the procedures in both cohorts were similar.

Incentives

Each team was promised a $4000 stipend for study completion (to be used for substitutes, team member stipends, or other professional learning expenses designated by the team). Each team was also promised a video camera to be used in data collection; the site could keep the camera with the stipulation that it be available for follow-up video-recording.

Eligibility criteria

To be eligible, at least three and at most eight teachers at a school site had to provide informed consent to participate, including at least one third- or fourth-grade teacher responsible for teaching mathematics. Only one team per school site was eligible. The district and school principal also had to approve the terms of the study. One of the consenting third- or fourth-grade mathematics teachers agreed to serve as the research-lesson teacher (RLT).Footnote1 RLTs also agreed to obtain parental consent for students in their own classrooms, teach a series of three fractions lessons to be video-recorded, and administer a pre- and postintervention fractions assessment to their students. The students of the other participating teachers were not eligible for the study. For sites assigned to the LS or LS + FR conditions, the RLT was also expected to teach the research lesson, which is a crucial part of the lesson study process, as one of the series of three videotaped lessons.

Consent process

The RLTs distributed parental consent and student assent forms to the students in their mathematics classes and mailed the returned forms to the research team with the student pretests and class rosters. The study included only those students who had both parental consent and student assent before the sites were notified of their assigned treatment condition.

Randomization

The study employed a randomized block design. The unit of randomization was the team of teachers who qualified for enrollment in the study based on the criteria listed above. Teams in eligible schools identified the RLT and corresponding class of students at the beginning of the school year (before randomization). Schools were grouped into blocks of four with the goal of achieving within-block homogeneity with respect to the following three criteria in descending order of importance: state, grade level of the designated RLT, and proportion of students eligible for free or reduced-price lunch. Every block could not be exactly homogenous with respect to those three variables. Each site in a given block was assigned at random to one of the four assignment conditions and notified of its treatment condition after consent for participation and baseline data were obtained for both teachers and students at that site.

Participants

presents school characteristics for the full sample, split by condition. Ten states and two grade levels of students are represented in the sample; 56% of the sites were in Florida. State, grade level, and the percentage of students in the school who were eligible for free or reduced-price lunch were relatively balanced across the four conditions.

Table 3. School characteristics.

presents information about all participating teachers. Only teachers who had consented to participate in the study at the time that their schools were randomized were included. The mean number of years of teaching experience was 13.5 (SD = 8.0). Sample characteristics were relatively balanced across the four conditions. The RLT sample was similar to the total sample of teachers in its composition, and the RLT sample was relatively balanced in characteristics across conditions. presents information about participating RLTs. presents information pertaining to the sample of participating students.

Table 4. Participating teacher characteristics.

Table 5. Research lesson teacher characteristics.

Table 6. Consenting student characteristics for sample.

Data collection

Teacher surveys and assessments

After at least three educators at a school had signed up and designated a third- or fourth-grade RLT, they were invited to complete the Web-based baseline assessments, which included knowledge assessments and demographic information. At the end of the year, the teachers in all four conditions completed a postintervention assessment, which again included the knowledge assessment and asked teachers to report on their experiences with professional learning over the past year.

School and student demographics

Demographic information for individual students with parental permission was provided by teachers on a class-roster form that was mailed to researchers along with student pretests, parental-consent, and student-assent forms.

Videos of classroom instruction

Each site was asked to videorecord the first three lessons of the fractions unit taught by the designated RLT, one of which was expected to be the designated research lesson in the LS + FR and LS conditions. Students without consent for videorecording were seated outside the camera view.

Student assessments

Student fractions assessments were administered by teachers in the fall and spring of their participating school year to students taught by the RLT. The same test forms were administered to both third- and fourth-grade students. Administered by teachers after the fractions unit was completed, the postintervention test form was different and more difficult than the pre-intervention form, but a set of anchor items were consistent on both test forms.

Measures

Teacher knowledge of subject matter

The computer-based Knowledge for Teaching Elementary Fractions (K-TEF) tests measure teacher knowledge of fractions at the elementary level. The test contains 19 items: 12 drawn from the Learning Mathematics for Teaching (Citation2004) project and 7 adapted from other sources (Beckmann, Citation2005; Newton, Citation2008; Saderholm et al., Citation2010; Schifter, Citation1998; Ward & Thomas, Citation2015; Zhou et al. Citation2006). Using response data for the K-TEF tests provided by 277 teachers in fall 2016 (i.e., Cohort 1 in this study), Schoen et al. (Citation2018b) found that the K-TEF measures a single, dominant factor. Coefficient α for the K-TEF pretest scores with this sample was .76.

Implementation

A full investigation of implementation is beyond the scope of the current study, but we provide some highlights of the implementation study in the main body of this article and more information in the Supplementary Online Materials. We focus on three facets of implementation, all three of which are central to the theory of change for the intervention: evidence of implementation of a research lesson, evidence of the use of the materials in the fractions resource kit, and evidence of the use of linear representations of fractions during fractions instruction.

Linear representations of fractions

A total of 181 fractions-lesson videos representing 68 sites were received (3 videos per site were requested), and all were coded to reveal whether lessons used linear representations of fractions. If any lesson in any video from the site used a linear fraction representation for at least one-half of the lesson, the site received a 1 for use of linear representations of fractions. Otherwise, the site was given a score of 0 on use of linear representations of fractions.

Evidence of a research lesson

We created a dichotomous variable for each site to indicate whether we observed either of the following conditions in the submitted videos of classroom instruction for each site: (a) two or more adult observers were present for the lesson or (b) at least one adult observer was focused on students and taking notes.

Mystery strip activity

The Mystery Strip activity comes from a Japanese textbook (Hironaka & Sugiyama, Citation2006) and was included in the fractions resources (both as a problem for teachers to try to solve and as a task featured in a set of lesson videos provided in the resources). The Mystery Strip activity is not widely known or used in the U.S., so its use in a submitted lesson video provides a strong indicator that the RLT looked at the provided fractions resources. Use of the Mystery Strip activity was coded from classroom videos by means of a dichotomous variable that recorded its use in any submitted lesson video (1 for presence, 0 for absence). Sites were not specifically asked to use the Mystery Strip activity in lessons. Therefore, we did not expect that all sites in the LS + FR or FR conditions would use the Mystery Strip activity in their videos, but teachers in the other two conditions (i.e., LS, C) should show no evidence of awareness of the Mystery Strip. Although presence of the Mystery Strip activity suggests the participant did use the fractions resources, absence of the Mystery Strip does not necessarily imply failure to use them.

Classroom instruction in fractions

Videos of fractions instruction were rated by means of the Instructional Quality Assessment (IQA; Boston, Citation2012). The IQA provided a treatment independent (Slavin & Madden, Citation2011) measure of classroom instruction, and it served as the primary outcome of interest for the confirmatory RQ1. Seven of the nine established IQA scales (i.e., Potential of the Task, Implementation of the Task, Student Discussion, Teacher Linking, Student Linking, Teacher Press, Student Providing) were used in scoring the videos. Each rubric was scored from zero to four. The first lesson video submitted by each of the 68 sites that returned videos was selected to be coded. The primary observer coded all 68 videos, and a second observer coded 15 of these. Discrepant scores between the two observers’ ratings were reconciled by the two observers and input as the final scores for those 15 videos. For all other videos, the primary observer’s ratings were used. We computed linear-weighted kappa (Cohen, Citation1968) using SPSS (version 27) for the two raters and found it to be 0.686 (p = .000). The arithmetic mean of the ratings on the seven rubrics was used as the overall IQA score.

Student Performance on fractions tests

Two versions of the Early Fractions Test (EFT; Schoen et al., Citation2017a, Citation2017b) were used—one version administered before the intervention, the other version administered after the intervention. Lewis and Perry (Citation2017) used an earlier version of the EFT in the initial efficacy study. Items on the EFT tests correspond to the grade-level expectations in the CCSSM (NGACBP & CCSSO, Citation2010) for third- and fourth-grade students, including linear representations of fractions (including the number line), referent unit, magnitude comparison, partitioning and iterating, and performing addition on fractions presented in symbolic notation.

Using response data for the EFT tests provided by 68 schools in Cohort 1, Schoen et al. (Citation2017a, Citation2017b) found that the EFT tests measure a single dominant factor. Coefficient α for the EFT pretest scores with the Cohort 1 sample was .85; the posttest coefficient α was found to be .84 with the Cohort 1 sample. Scoring procedures used a two-parameter logistic model based on item-response theory, scores were not vertically scaled across pre- and posttest waves of data collection, and scores were generated by the expected a priori estimator (Schoen et al., Citation2017a, Citation2017b).

Data analysis

Attrition analysis

The Institute of Education Science’s What Works Clearinghouse (WWC) publishes standards for determining whether the combination of overall and differential attrition rates is likely to bias estimates of average causal effects from randomized experiments (U.S. Department of Education, Citation2017). We provide an attrition analysis in the Supplementary Online Materials.

RQ1: Effects on classroom instruction

These analyses use data from the 68 RLT teachers who submitted at least one video of fractions instruction in their classrooms. The model for the IQA outcome was an ordinary least squares regression model with FLORIDA, PCTFRL, the school average student pretest value, a dummy variable indicating that most of the students in the RLT teacher’s classroom were in fourth grade, treatment dummy variables, and block dummy variables. The effect size for IQA was computed based on the residual standard error from an ordinary least squares model where the only independent variables are dummies for treatment status and cohort.

RQ2: Effects on Student achievement in fractions

This analysis used data from the 1,199 students who provided complete pre- and postintervention test data for the (EFT). A two-level hierarchical linear model with students nested within schools was estimated. Specifically, we estimated models with the following specifications:

Level 1: Yjk = τ0k + τ1k (PREC) jk +u0jk.

Level 2: τ0k = γ00 + γ01 (PREM)k + γ02 (FLORIDA)k + γ03 (PCTFRL)k + γ02 (GRADE4)k + Σγ0T (TREATT)k + Σγ0M(BLOCKM)k + r00k

τ1k = γ10

Yjk represents the EFT student postintervention test score, PREC is the school-mean-centered EFT student preintervention test score, PREM is the school mean of the EFT preintervention test score, FLORIDA is an indicator variable equal to 1 if the school is located in Florida and 0 otherwise, PCTFRL is a continuous variable equal to the percentage of students eligible for free or reduced-price lunch at the school, GRADE4 is an indicator variable equal to 1 if the grade level of the RLT classroom was composed of fourth-grade students, TREATT (T = 1, …, 3) are dummy variables for the three active treatment conditions (LS + FR, LS, and FR), and BLOCKM (M = 1,…, 19) are dummy variables for the randomization blocks.

Effect sizes associated with each active treatment condition are reported as the coefficients for the treatment dummies divided by the unconditional standard deviation of the outcome variable of interest. The unconditional standard deviation is computed as the sum of the square root of the two estimated variance components from a two-level model where the only independent variables are dummies for treatment status and cohort.

RQ3: Moderation by student pretest, teacher pretest, grade level, and school location

This research question addresses moderating effects of the four variables listed relative to each of the outcome variables considered in RQ1 (classroom instruction) and RQ2 (student achievement). The continuous moderators (student pretest and site mean for K-TEF pretest) were z-scored before entering statistical models. The student grade moderator was coded 1 if most of the students in the RLT’s classroom were in fourth grade and 0 otherwise. The school-location moderator was coded 1 if the school was in the state of Florida and 0 otherwise. (Approximately half of the sites were in Florida.) The binary moderators were not z-scored.

For the Florida variable, site-mean K-TEF score, and RLT grade-level moderators, the moderating variable was defined in the same way regardless of whether IQA or EFT measured the outcome. For the student preintervention test EFT moderator, the individual student test was used as the moderator when EFT was the outcome, and the site-mean student test was used when IQA was the outcome.

Moderator models were specified by elaboration of the models used to answer research questions 1 and 2. Specifically, interactions between the moderator in question and each of the three dummy treatment variables were included in the models. Otherwise, the statistical models used were unchanged.

Results

Implementation of the intervention components

While not a research question, implementation of an intervention, including the implementation and experiences of participants in the counterfactual condition during the intervention period, is an important consideration for interpreting results of a randomized controlled trial. presents results of our three implementation indicators for each treatment condition overall and disaggregated by whether the site was in Florida. A more complete description of our implementation analyses can be found in the Supplemental Online Materials.

Table 7. Implementation activities observed in videos (by location and randomization status).

Linear representations of fractions were observed in videos of fractions instruction in all four conditions. It was observed in a higher proportion of videos from the FR and LS + FR conditions than in the LS or C conditions. We did not perform tests to examine whether these differences are statistically significant, because we did not design the study of implementation for inferential statistics. We note that the point estimate for implementation rates of linear representations of fractions instruction is lowest in the condition that conducted lesson study without the infusion of fractions resources and highest in the condition that was provided fractions resources and asked teachers to study them on their own (i.e., not in collaboration with the others at their site).

We found a lower-than-expected rate of implementation of research lessons in the LS and LS + FR conditions and a higher-than expected rate in the other two conditions, especially in the C condition. This suggests low implementation rate in the two conditions involving lesson study and a potential contamination of the C condition, especially in sites located outside of Florida.

The Mystery Strip activity was not observed in the two conditions that did not receive the activity as part of the intervention. We did observe the activity being used in videos of fractions instruction in the two conditions that received that as part of the fractions resources. We remind the reader that—unlike the research lesson—those groups were not specifically asked to submit videos of the Mystery Strip lesson in use. The presence of the activity in some of the videos does suggest that the sites used the fractions resources. The absence of the activity in the LS and C conditions suggests minimal diffusion of the fractions resources into the LS or C conditions.

RQ1: Effects on fractions instruction

The first row of shows estimates of the effects of each of the three treatment conditions on fractions instruction as compared with the comparison condition (C). Fractions instruction at sites assigned to both the conditions that received fraction resources scored significantly higher on the IQA than that of sites in the comparison condition. The effect-size estimate was also positive for the LS condition, but the magnitude was much smaller, and the p-value was much greater than .05 for that comparison. The first row of shows the raw means and standard deviations for the IQA measure, separately for each treatment group.

Table 8. Treatment contrasts for fractions instruction and student learning.

Table 9. Mean, standard deviation, and sample size by treatment condition for the analytic samples used to answer research questions 2 and 3.

RQ2: Effects on student performance on the fractions test

The last row of shows that students taught by comparison-condition teachers performed best on the EFT, but none of the pairwise differences were statistically significant. The effect size where the LS + FR group is compared to control was estimated to be −0.075, that where the LS group is compared to control was −0.12, and that where the FR group is compared to control was −0.19. The last two rows of show the raw means and standard deviations for the EFT measure at pre- and post-intervention test, separately by treatment group. Overall, there is little evidence of stable differences among the experimental conditions with respect to student performance on the EFT.

RQ3: Moderator effects

We present the results of moderator analyses in , which shows uncorrected p-values and four moderators for three pairwise comparisons of interest. For each outcome variable, therefore, 12 moderation effects are estimated. This multiplicity problem makes it difficult to distinguish signal from noise in our moderator analyses, because the magnitudes of certain effects will appear different from those of others just by chance, due to the number of effects estimated. If a Bonferroni correction was applied to account for the 12 comparisons, none of the results would be statistically significant at the conventional 0.05 level. Even if the more liberal Benjamini-Hochberg (Citation1995) procedure was used, one would not reject any of the null hypotheses unless the false discovery rate was set to a high value of 0.27 (that is to say, the lowest q value is 0.27). Nonetheless, the moderation analyses do present some suggestive results.

Table 10. Moderator analyses with instructional quality assessment and early fractions test as outcomes.

We first discuss moderator analyses where student performance on the EFT was the outcome. There is no evidence to suggest that treatment effects vary systematically with respect to student pretest scores. Although not statistically significant, some evidence suggests that the negative contrast between FR and C students was mainly due to negative effects in third grade. The fitted means for fourth-grade students in the FR and C groups are, in fact, almost identical. Some evidence also suggests that, among the sites in the LS condition, the average teacher knowledge of fractions of the lesson study team is negatively related to treatment effects, suggesting that, relative to similarly situated control group students, students performed comparatively worse on the test if they had high-knowledge teachers who were asked to engage in lesson study without the support of the fractions resources.

The most striking results for the EFT outcome involved the moderating effect of state. Evidence suggests substantial differences in the impact of lesson study between Florida and non-Florida sites. For schools not in Florida, the estimate of the impact of both LS and LS + FR relative to control was negative and statistically significant (without multiplicity corrections). The estimated effect sizes were −0.58 for LS and −0.48 for LS + FR for non-Florida schools. The effect estimated for the FR condition relative to control was also negative, but it was smaller and not statistically significant. On the other hand, a large positive interaction was estimated for the moderating effect of a location in Florida. For both LS and LS + FR, the moderating effect of Florida is positive, statistically significant, and greater in absolute value than the estimated negative impact in states other than Florida. In other words, evidence suggests that impacts of lesson study were, on average, negative outside of Florida but positive within Florida (although the positive-within-Florida estimates are not statistically significant). The same pattern of positive moderation was observed for the FR condition, but it was a much smaller effect. Even after this positive moderating effect was accounted for, the average treatment effect estimate for FR vs. C within Florida was negative.

Some evidence also indicates heterogeneity in effects across certain moderators for the classroom-instruction outcome. Like the student outcome, it is not moderated by the (average) student pretest. Some evidence indicates a positive treatment effect on IQA in the LS group for fourth-grade teachers and a negative effect for third-grade teachers, but the result is not statistically significant, even without multiplicity corrections. Slightly more evidence indicates moderation with respect to the average level of fractions knowledge in the participating teacher group. Moderation again seems to occur only for the LS group; outcomes were worse (relative to the comparison condition) for groups with a higher average level of fractions knowledge.

Again, the most striking results came from examination of state as a moderator. For sites in states other than Florida, a small, not statistically significant, negative impact of treatment on IQA was evident for the LS and LS + FR groups, but state had a strong moderating effect. Within Florida, effect size estimates were positive, statistically significant, and greater than 0.5 for the LS, LS + FR and the FR conditions.

Discussion

Impact of fractions resources on instruction

A statistically significant (p < .05) effect on classroom instruction as measured by the IQA was found in the two conditions that received fractions resources, but a similar effect was not found in the condition assigned to lesson study alone. The resources provided to the sites in the two fractions resources conditions are more extensive than those U.S. teachers typically receive as part of curriculum adoption (Remillard, Citation2000); they included research-based summaries of students’ fractions challenges, fractions tasks, curriculum materials highlighting linear representations of fractions, video of fractions instruction, and prompts to discuss (or reflect on) all of these. The positive impact on instruction in the two conditions that received the fractions resources provides evidence that mathematical resources and the opportunity to study them—whether individually or collaboratively—can increase quality of instruction.

Effects on student learning

The average effects of each of the three intervention conditions on student performance on fractions tests were negative and close to zero in all three cases. Additional explorations are needed to reveal the reason for these overall null effects, which contrast with findings of a prior study (Lewis & Perry, Citation2017) of a similar intervention with a similar outcome measure.

The positive effects on IQA (for the two conditions that received fractions resources) and negative effects on student achievement are perplexing and raise questions about the association between IQA scores and student learning. The analyses are not presented here, but we sought to determine whether IQA scores predict student learning, and we found a small, positive, statistically significant (α < .05) correlation between IQA scores and student performance on the postintervention EFT when adjusted for the preintervention performance (Authors, manuscript under review). Many researchers in mathematics education are using IQA in their studies. Only a small subset of them have reported on empirical associations between IQA scores and student achievement in mathematics, but those that have reported on the association report a positive association (Lewis et al., manuscript under review; Matsumura et al., Citation2008; Quint et al., Citation2007). However, the association between IQA and student performance found in the current study may be driven by a confounding factor that we have not yet identified. Our results underscore the need for further empirical study to identify the measurable components that lead to both improved instruction and increased student learning in fractions and the causal links between the two.

Moderators of intervention impact

None of the four moderators—student fractions pretest, grade level, site mean of teachers’ mathematical knowledge, and state (i.e., Florida, non-Florida)—showed significant impact after correction for multiplicity. The moderator analyses do, however, suggest positive impact of the two lesson-study conditions on instructional and student outcomes in Florida and negative average impacts in non-Florida sites.

To further explore explanations for the treatment-by-Florida interaction, we split the implementation data into Florida and non-Florida sites (). We found that all four of the C condition sites that presented evidence of a research lesson (i.e., that failed to abide by the control-condition restrictions) were outside Florida. A lack of adherence to the assigned condition could explain both the Florida moderator and the apparent discrepancy between the findings of the present study and those of the previous randomized controlled trial of lesson study with fractions resources. We note that the Florida sites were recruited through a general call for research-study participants that was sent by the first author to district mathematics curriculum supervisors and principals across the state. The sites outside Florida were recruited by the second author through lesson study networks and professional learning networks. This difference in recruiting method may explain why sites outside of Florida more frequently engaged in lesson-study-type activities, even when asked to refrain from doing so. In this study, the location of sites (within or outside Florida) and the method of recruiting sites is confounded. So we cannot be sure whether the state context, the recruitment method, or some combination of the two produced the differences between sites within and outside Florida.

Limitations

The present study included teams of educators who volunteered, limiting generalizability to non-volunteers. The marginal levels of fidelity to the intervention, especially in the LS + FR condition, and lack of a sharp contrast across conditions make inferences about the underlying program theory of change difficult. The three videorecorded lessons were not selected at random from the full unit on fractions and may not represent the full experience of students in fractions, possibly explaining the positive effects on classroom instruction and null effects on student learning. The improvement in instructional quality in the LS + FR and FR conditions might not extend to other topics in mathematics or to other grade levels, because the materials were specific to early fractions.

Conclusions

Although assignment to lesson study alone did not improve instruction, assignment to the two conditions that received mathematical resources (with and without lesson study) did produce significantly higher IQA scores. This result contributes to a growing collection of findings that professional-development opportunities that are, by design, combined with curriculum resources or other actionable materials (e.g., lesson plans, tasks to be used in formative assessment) may be more likely to have positive effects on teaching and learning (Garet et al., Citation2016; Harris et al., Citation2015; Jacob et al., Citation2017; Lang et al., Citation2014; Lonigan et al., Citation2011; Penuel et al., Citation2011; Rosenfeld et al., Citation2019; Schoen & Koon, Citation2021). These apparent improvements in instruction, however, did not lead to improvements in student performance on fractions tests in this study.

We observed lower-than-expected implementation of lesson study in the two lesson-study conditions, substantial use of linear fractions representations in all conditions, and substantial collaboration in all conditions, including some lesson-study-like activities in the control condition. Binders and guidance about expected and impermissible activities failed to produce the expected contrast across conditions. We imagined that providing the fractions standards in the control condition binder would have little impact, since these standards are readily available. Clearer guidance and establishment of mutual expectations for the activities in the C condition might have yielded a different outcome. Implementation data suggest that some sites/teachers may need more, or different, support for implementation of the materials, especially in the LS + FR condition.

Evidence of a treatment-by-state interaction in the present study provides a reminder that context may matter. This program was designed to support local adaptation and agency, and it appears to have had varied results in different settings or with participants recruited in different ways. Another follow-up study may be warranted to increase fidelity of implementation of the designed programs and better reveal the conditions under which the interventions can be expected to have the intended effects.

Supplemental material

Supplemental Material

Download MS Word (17.5 KB)

Supplemental Material

Download MS Word (18.6 KB)

Supplemental Material

Download MS Word (17.6 KB)

Supplemental Material

Download MS Word (18.6 KB)

Disclosure statement

The opinions expressed are those of the authors and do not represent views of the Institute or the U.S. Department of Education.

Data availability statement

The data that support the findings of this study are openly available via ICPSR at https://doi.org/10.3886/ICPSR119304.v1 [ICPSR 37080].

Additional information

Funding

The research reported here was supported by the Institute of Education Sciences, U.S. Department of Education, through Grant # R305A150043 to Mills College.

Notes

1 While no research lessons were taught in sites assigned to the FR or C conditions we still refer to teachers performing the tasks described in the next sentences as “RLTs.”

References

  • Beckmann, S. (2005). Mathematics for elementary teachers. Addison-Wesley.
  • Benjamini, Y., & Hochberg, Y. (1995). Controlling the false discovery rate: A practical and powerful approach to multiple testing. Journal of the Royal Statistical Society Series B, 57(1), 289–300.
  • Booth, J. L., & Newton, K. J. (2012). Fractions: Could they really be the gatekeeper’s doorman? Contemporary Educational Psychology, 37(4), 247–253. https://doi.org/10.1016/j.cedpsych.2012.07.001
  • Boston, M. D. (2012). Assessing instructional quality in mathematics. The Elementary School Journal, 113(1), 76–104. https://doi.org/10.1086/666387
  • Cohen, D. K. (2011). Teaching and its predicaments. Harvard University Press.
  • Cohen, J. (1968). Weighted kappa: Nominal scale agreement with provision for scaled disagreement or partial credit. Psychological Bulletin, 70(4), 213–220. https://doi.org/10.1037/h0026256
  • Common Core Standards Writing Team. (2013). Progressions for the common core state standards in mathematics: Fractions. Institute for Mathematics and Education, University of Arizona.
  • Doig, B., Groves, S., & Fujii, T. (2011). The critical role of task development in Lesson Study. In L. C. Hart, A. Alston, & A. Murata (Eds.), Lesson Study research and practice in mathematics education (pp. 181–199). Springer.
  • Fermanich, M. L. (2002). School spending for professional development: A cross-case analysis of seven schools in one urban district. The Elementary School Journal, 103(1), 27–50. https://doi.org/10.1086/499714
  • Fuchs, L. S., Malone, A. S., Schumacher, R. F., Namkung, J., & Wang, A. (2017). Fraction intervention for students with mathematics difficulties: Lessons learned from five randomized controlled trials. Journal of Learning Disabilities, 50(6), 631–639. https://doi.org/10.1177/0022219416677249
  • Fujii, T. (2014). Implementing Japanese Lesson Study in foreign countries: Misconceptions revealed. Mathematics Teacher Education and Development, 16(1), 65–83.
  • Garet, M. S., Heppen, J. B., Walters, K., Parkinson, J., Smith, T. T. M., Song, M. (2016). Focusing on mathematical knowledge: The impact of content-intensive teacher professional development (NCEE 2016-4010). U.S. Department of Education, Institute of Education Sciences, National Center for Education Evaluation and Regional Assistance. http://ies.ed.gov/ncee/pubs/20164010/pdf/20164010.pdf
  • Gearhart, M., Saxe, G. B., Seltzer, M., Schlackman, J., Ching, C. C., Nasir, N., Fall, R., Bennett, T., Rhine, S., & Sloan, T. F. (1999). Opportunities to learn fractions in elementary mathematics classrooms. Journal for Research in Mathematics Education, 30(3), 286–315. https://doi.org/10.2307/749837
  • Gersten, R., Taylor, M. J., Keys, T. D., Rolfhus, E., Newman-Gonchar, R. (2014). Summary of research on the effectiveness of math professional development approaches (REL 2014-010). U.S. Department of Education, Institute of Education Sciences, National Center for Educational Evaluation and Regional Assistance, Regional Educational Laboratory Southeast. http://ies.ed.gov/ncee/edlabs
  • Grow-Maienza, J., & Beal, S. (2005, April 4–6). What we can learn from Asian mathematics textbooks. Presented at the Research Presession of the National Council of Teachers of Mathematics.
  • Harris, C. J., Penuel, W. R., D'Angelo, C. M., DeBarger, A. H., Gallagher, L. P., Kennedy, C. A., Cheng, B. H., & Krajcik, J. S. (2015). Impact of project-based curriculum materials on student learning in science: Results of a randomized controlled trial. Journal of Research in Science Teaching, 52(10), 1362–1385. https://doi.org/10.1002/tea.21263
  • Hill, H. C., Beisiegel, M., & Jacob, R. (2013). Professional development research: Consensus, crossroads, and challenges. Educational Researcher, 42(9), 476–487. https://doi.org/10.3102/0013189X13512674
  • Hironaka, H., & Sugiyama, Y. (2006). Mathematics for Elementary School, Grades 1–6 (A. Takahashi, T. Watanabe, J. Shouffler, & M. Leer, Trans.). Tokyo Shoseki Co., Ltd. (English language version available at www.globaledresources.com).
  • Jacob, R., Hill, H., & Corey, D. (2017). The impact of a professional development program on teachers’ mathematical knowledge for teaching, instruction, and student achievement. Journal of Research on Educational Effectiveness, 10(2), 379–407. https://doi.org/10.1080/19345747.2016.1273411
  • Jayanthi, M., Gersten, R., Schumacher, R. F., Dimino, J., Smolkowski, K., & Spallone, S. (2021). Improving struggling fifth-grade students’ understanding of fractions: A randomized controlled trial of an intervention that stresses both concepts and procedures. Exceptional Children, 88(1), 81–100. https://doi.org/10.1177/00144029211008851
  • Lamon, S. J. (2005). Teaching fractions and ratios for understanding: Essential content knowledge and instructional strategies for teachers. Lawrence Erlbaum Associates.
  • Lang, L. B., Schoen, R. C., LaVenia, M., Oberlin, M. (2014, March). Mathematics Formative Assessment System–Common Core State Standards: A randomized field trial in kindergarten and first grade. Society for Research in Educational Effectiveness Annual Spring Conference, Washington, DC.
  • Learning Mathematics for Teaching. (2004). Mathematical knowledge for teaching measures: Geometry content knowledge, number concepts and operations content knowledge, and patterns and algebra content knowledge. Author.
  • Lee, S., & Zusho, A. (2002). Comparing Japanese and U.S. teachers’ manuals: Implications for mathematics teaching and learning. In G. DeCoker (Ed.), National standards and school reform in Japan and the United States. Teachers College Press.
  • Lewis, C., Friedkin, S., Emerson, K., Henn, L., & Goldsmith, L. (2019). How does Lesson Study work? Toward a theory of Lesson Study process and impact. In R. Huang, A. Takahashi, & J. P. da Ponte (Eds.), Theory and practice of lesson study in mathematics (pp. 13–37). Springer.
  • Lewis, C., & Hurd, J. (2011). Lesson study step by step: How teacher learning communities improve instruction. Heinemann.
  • Lewis, C. C., & Perry, R. (2017). Lesson study to scale up research-based knowledge: A randomized-controlled trial of fractions learning. Journal for Research in Mathematics Education, 48(3), 261–299. https://doi.org/10.5951/jresematheduc.48.3.0261
  • Lewis, C., Perry, R., & Friedkin, S. (2011). Using Japanese curriculum materials to support Lesson Study outside Japan: Toward coherent curriculum. Educational Studies in Japan, 6, 5–19. https://doi.org/10.7571/esjkyoiku.6.5
  • Lonigan, C. J., Farver, J. M., Phillips, B. M., & Clancy-Menchetti, J. (2011). Promoting the development of preschool children’s emergent literacy skills: A randomized evaluation of a literacy-focused curriculum and two professional development models. Reading and Writing, 24(3), 305–337. https://doi.org/10.1007/s11145-009-9214-6
  • Ma, L. (1999). Knowing and teaching elementary mathematics. Lawrence Erlbaum Associates.
  • Matsumura, L. C., Garnier, H. E., Slater, S. C., & Boston, M. D. (2008). Toward measuring instructional interactions “at-scale. Educational Assessment, 13(4), 267–300. https://doi.org/10.1080/10627190802602541
  • Miyakawa, T. (2011). Textbooks and teaching guides. In M. Isoda, M. Stephens, Y. Ohara, & T. Miyakawa (Eds.), Japanese Lesson Study in mathematics: Its impact, diversity, and potential for educational improvement (pp. 48–51). World Scientific.
  • Morris, A. K. (2000). A teaching experiment: Introducing fourth graders to fractions from the viewpoint of measuring quantities using Davydov’s mathematics curriculum. Focus on Learning Problems in Mathematics, 22(2), 32–84.
  • National Mathematics Advisory Panel. (2008). Foundations for success: The final report of the National Mathematics Advisory Panel. U.S. Department of Education.
  • National Governors Association Center for Best Practices & Council of Chief State School Officers. (2010). Common core state standards for mathematics. Authors.
  • Newton, K. J. (2008). An extensive analysis of preservice elementary teachers’ knowledge of fractions. American Educational Research Journal, 45(4), 1080–1110. https://doi.org/10.3102/0002831208320851
  • Odden, A., Archibald, S., Fermanich, M., & Gallagher, H. A. (2002). A cost framework for professional development. Journal of Education Finance, 28(1), 51–74.
  • OECD (Organisation for Economic Co-operation and Development). (2014). A profile of student performance in mathematics. In PISA 2012 Results: What Students Know and Can Do (Volume I, Revised edition, February 2014): Student Performance in Mathematics, Reading and Science (pp. 31–144). OECD Publishing. https://doi.org/10.1787/9789264208780-6-en
  • Penuel, W., Gallagher, L., & Moorthy, S. (2011). Preparing teachers to design sequences of instruction in earth systems science: A comparison of three professional development programs. American Educational Research Journal, 48(4), 996–1025. https://doi.org/10.3102/0002831211410864
  • Quint, J. C., Akey, T. M., Rappaport, S., & Willner, C. J. (2007). Instructional leadership, teaching quality and student achievement: Suggestive evidence from three urban school districts. MDRC.
  • Remillard, J. T. (2000). Can curriculum materials support teachers’ learning? Two fourth-grade teachers’ use of a new mathematics text. The Elementary School Journal, 100(4), 331–350. https://doi.org/10.1086/499645
  • Rosenfeld, D., Dominguez, X., Llorente, C., Pasnik, S., Moorthy, S., Hupert, N., Gerard, S., & Vidiksis, R. (2019). A curriculum supplement that integrates transmedia to promote early math learning: A randomized controlled trial of a PBS KIDS intervention. Early Childhood Research Quarterly, 49, 241–253. https://doi.org/10.1016/j.ecresq.2019.07.003
  • Saderholm, J., Ronau, R., Brown, E. T., & Collins, G. (2010). Validation of the Diagnostic Teacher Assessment of Mathematics and Science (DTAMS) instrument. School Science and Mathematics, 110(4), 180–192. https://doi.org/10.1111/j.1949-8594.2010.00021.x
  • Saxe, G. B. (2007). Learning about fractions as points on a number line. In W. G. Martin, M. E. Strutchens, & P. C. Elliott (Eds.), The learning of mathematics: Vol. 69th yearbook (pp. 221–237). The National Council of Teachers of Mathematics.
  • Saxe, G. B., Diakow, R., & Gearhart, M. (2013). Towards curricular coherence in integers and fractions: A study of the efficacy of a lesson sequence that uses the number line as the principal representational context. ZDM, 45(3), 343–364. https://doi.org/10.1007/s11858-012-0466-2
  • Saxe, G. B., Gearhart, M., & Nasir, N. (2001). Enhancing students’ understanding of mathematics: A study of three contrasting approaches to professional support. Journal of Mathematics Teacher Education, 4(1), 55–79. https://doi.org/10.1023/A:1009935100676
  • Schifter, D. (1998). Learning mathematics for teaching: From a teachers’ seminar to the classroom. Journal of Mathematics Teacher Education, 1(1), 55–87. https://doi.org/10.1023/A:1009911200478
  • Schmidt, S. (2009). Shall we really do it again? The powerful concept of replication is neglected in the social sciences. Review of General Psychology, 13(2), 90–100. https://doi.org/10.1037/a0015108
  • Schoen, R. C., Koon, S. (2021). The effect of a curriculum and professional-development program on middle-school students’ understanding of statistics (REL 2021-055). U.S. Department of Education, Institute of Education Sciences, National Center for Education Evaluation and Regional Assistance, Regional Educational Laboratory Southeast. https://nces.ed.gov/pubsearch/pubsinfo.asp?pubid=REL2021055
  • Schoen, R. C., LaVenia, M., & Tazaz, A. M. (2018a). Effects of the first year of a three-year CGI teacher professional development program on grades 3–5 student achievement: A multisite cluster-randomized trial (Research Report No. 2018-25). Florida State University. https://doi.org/10.33009/fsu.1562595733
  • Schoen, R. C., Liu, S., Yang, X., & Paek, I. (2017a). Psychometric report for the Early Fractions Test administered with third- and fourth-grade students in fall 2016 (Research Report No. 2017-10). Florida State University. https://doi.org/10.17125/fsu.1512509662
  • Schoen, R. C., Yang, X., Liu, S., & Paek, I. (2017b). Psychometric report for the Early Fractions Test (version 2.2) administered with third- and fourth-grade students in spring 2017 (Research Report No. 2017-11). Florida State University. https://doi.org/10.17125/fsu.1522698235
  • Schoen, R. C., Yang, X., Liu, S., & Paek, I. (2018b). Psychometric report on the Knowledge for Teaching Elementary Fractions Test administered to elementary educators in six states in fall 2016 (Research Report No. 2018-12). Florida State University. https://doi.org/10.17125/fsu.1531453537
  • Sidney, P. G., Thompson, C. A., & Rivera, F. D. (2019). Number lines, but not area models, support children’s accuracy and conceptual models of fraction division. Contemporary Educational Psychology, 58, 288–298. https://doi.org/10.1016/j.cedpsych.2019.03.011
  • Siegler, R. S., Carpenter, T., Fennell, F., Geary, D., Lewis, J., Okamoto, Y., Thompson, L., Wray, J. (2010). Developing effective fractions instruction for kindergarten through 8th grade: A practice guide ((NCEE #2010-4039); IES Practice Guide). National Center for Education Evaluation and Regional Assistance, Institute of Education Sciences, U.S. Department of Education. http://ies.ed.gov/ncee/wwc/pdf/practice_guides/fractions_pg_093010.pdf
  • Siegler, R. S., Fazio, L. K., Bailey, D. H., & Zhou, X. (2013). Fractions: The new frontier for theories of numerical development. Trends in Cognitive Sciences, 17(1), 13–19. https://doi.org/10.1016/j.tics.2012.11.004
  • Slavin, R., & Madden, N. A. (2011). Measures inherent to treatments in program effectiveness reviews. Journal of Research on Educational Effectiveness, 4(4), 370–380. https://doi.org/10.1080/19345747.2011.558986
  • Takahashi, A. (2021). Teaching mathematics through problem-solving: A pedagogical approach from Japan. Routledge.
  • Takahashi, A., & McDougal, T. (2014). Implementing a new national curriculum: Case study of a Japanese school’s 2-year lesson-study project. In K. Karp (Ed.), Annual perspectives in mathematics education: Using research to improve instruction 2014. National Council of Teachers of Mathematics.
  • Takahashi, A., Watanabe, T., Yoshida, M., & Wang-Iverson, P. (2005). Improving content and pedagogical knowledge through kyozaikenkyu. In P. Wang-Iverson & M. Yoshida (Eds.), Building our understanding of lesson study (pp. 77–84). Research for Better Schools.
  • TNTP (The New Teacher Project). (2015). The mirage: Confronting the hard truth about our quest for teacher development. Author. https://tntp.org/assets/documents/TNTP-Mirage_2015.pdf
  • Tokyo Shoseki Co. (2015). Sansu Math: Vol. Grade 2. Koyo Publishing Inc.
  • U.S. Department of Education. (2014). Fiscal year 2014 budget summary and background information. Washington, DC. http://www2.ed.gov/about/overview/budget/budget14/summary/14summary.pdf
  • U.S. Department of Education. (2017). What works clearinghouse standards handbook version 4.0. Institute of Education Sciences, What Works Clearinghouse.
  • Ward, J., & Thomas, G. (2006). What do teachers know about fractions? Findings from the New Zealand Numeracy Development Projects 2006. Ministry of Education.
  • Ward, J., & Thomas, G. (2015). Numeracy development project. Retrieved October 14, 2017, from http://www.nzmaths.co.nz/sites/default/files/Numeracy/FractionsScenarios.pdf
  • Watanabe, T. (2002). Representations in teaching and learning fractions. Teaching Children Mathematics, 8(8), 457–463. https://doi.org/10.5951/TCM.8.8.0457
  • Watanabe, T. (2007). Initial treatment of fractions in Japanese textbooks. Focus on Learning Problems in Mathematics, 29(2), 41–60.
  • Wei, R. C., Darling-Hammond, L., & Adamson, F. (2010). Professional development in the United States: Trends and challenges. National Staff Development Council.
  • Wu, H. H. (2011). Understanding numbers in elementary school mathematics. American Mathematical Society.
  • Yoon, K. S., Duncan, T., Lee, S. W.-Y., Scarloss, B., & Shapley, K. (2007). Reviewing the evidence on how teacher professional development affects student achievement (Issues and Answers Report, REL 2007-No. 033). U.S. Department of Education, Institute of Education Sciences, National Center for Education Evaluation and Regional Educational Laboratory Southwest. http://ies.ed.gov/ncee/edlabs
  • Yoshida, M., & Jackson, W. C. (2011). Response to Part V: Ideas for developing mathematical pedagogical content knowledge through Lesson Study. In L. C. Hart, A. Alston, & A. Murata (Eds.), Lesson study research and practice in mathematics education. Springer.
  • Zhou, Z., Peverly, S. T., & Xin, T. (2006). Knowing and teaching fractions: A cross-cultural study of American and Chinese mathematics teachers. Contemporary Educational Psychology, 31(4), 438–457. https://doi.org/10.1016/j.cedpsych.2006.02.001