Article

The Terhar debacle roundup

Fallout from State Board of Education President's offensive and ignorant Facebook post, comparing President Obama to Hilter spread far and wide yesterday.

The latest development has a tepid statement from the Governor's spokesperson saying "the governor has no comment about Terhar’s posting and no plan to remove her as board president". A true profile in courage and leadership.

The Ohio Department of education had its head resign over a serious ethics violation, and the current acting superintendent and his deputy are reportedly applying for district level jobs. Education leadership in Ohio is in disarray. Now the governor expects education professionals to take the State Board of Education President seriously after this disgraceful incident? Certainly her fellow board members don't think so

“I don’t think she can be perceived as being objective,” said Jeffrey Mims, a board member from Dayton. “Whether she wrote the statement or not, the fact that she harbored those statements on her Facebook page, I think you just can’t be associated with that.”

Ann Jacobs, a board member from Lima, said the board is supposed to be nonpartisan but Terhar’s posting is the latest sign of how partisan it’s gotten in recent years.

Here's a sampling of some of the coverage

Her continued presence on the board brings disgrace to the State Board of Education, Ohio's education system and all the professionals who strive everyday to bring excellence to educating Ohio's young people.

HB555 FAQ for teacher evaluations

We've just got our hands on this document put out by the Ohio Department of Education. It's an updated framework for teacher evaluations based on the changes that were slipped into HB555 in the dead of night.

Is there a more ridiculously convoluted and complex framework for evaluating any other professions job performance? How is any teacher expected to understand all this enough to know where to focus improvement efforts, especially since the Value-added formula itself is secret and proprietary.

HB 555 FAQ with regard to teacher evaluations by

10 Education Reform Tactics That Hurt Students and Don’t Improve Education

We write almost exclusively about education reform here at JTF, and there's been an awful lot to write about in recent years. At the very core of our support and objections to various reforms has always been whats best for students. This post from LAProgressive captures a lot of the problems with the current direction corporate education reform is taking us, and the negative effect it has on students

1. Deluging schools with tests in every grade and every subject beginning with pre-kindergarten, to the point where little else goes on in school but preparing for tests.

2. Pushing the arts out of a central role in the life and culture of public schools.

3. Demoralizing teachers, especially the most talented and experienced teachers, by subjecting them to evaluations based on junk science

4. Discriminating against special needs and English Language Learner (ELL) students by giving favorable treatment to charter schools which exclude or drive out such students, and forcing such students to take tests that are developmentally inappropriate for them

5. Destabilizing communities by closing schools that have been important community institutions for generations.

6. Undermining the mentoring and relationship building that are at the core of great teaching, especially in poor and working class communities, by raising class size and substituting online learning for direct instruction without thinking through the consequences of such policies on young people who need personal attention and guidance.

7. Creating such unrealistic pressure on schools, and on administrators and teaching staffs, that cheating on tests becomes endemic.

8. Giving billionaire philanthropists, and wealthy companies which provide services to schools such power over education policy smothering the voices of teachers, parents and students.

9. Replacing veteran teachers, often teachers of color, with poorly trained Teach for America Corps members,most of them white, who go through a 5 week training period before being given their own class, and often leave for other professions after their two year teaching commitment is completed.

10. Adding to mental health problems of students by spending so much on testing that school districts have to fire school counselors, and to the physical problems of students by transforming gym and recess and after school programs into test prep removing opportunities for exercise and play.

There's a lot to recognize in that list, and be worried about.

Debe Terhar breaks Godwin's law

The State Board of Education President Debe Terhar, in all her professional glory

State Board of Education President Debe Terhar said she was not comparing President Barack Obama to Adolf Hitler when she posted a photograph of the Nazi leader on her Facebook page with a message critical of the administration’s new gun-control efforts.

But she does say we “need to step back and think about it and look at history” to see that tyrants have disarmed their citizens.

Terhar, a Cincinnati Republican elected last week by the 19-member school board to a second term as its president, recently posted the picture with this commentary: “Never forget what this tyrant said: ‘To conquer a nation, first disarm its citizens.’ — Adolf Hitler.”

The photograph apparently originated with the Facebook page of Uncle Sam’s Misguided Children, which features a variety of anti-Obama, pro-gun posts and photos, such as scantily clad women hoisting large guns, a polar bear with the words “Holy f*** I’m glad I’m white,” and another saying “Where’s Lee Harvey Oswalt when you need him?”

When the Sipatch asked her about her Facebook post she said, “I’m not comparing the president to Adolf Hitler, it’s the thought of disarming citizens, and this has happened throughout history. What’s the true intention of the Second Amendment? It was to protect us from a tyrannical government, God forbid.”

Why is the State Board of Education President trolling racist Facebook pages in the first place, let alone making offensive comparisons to a genocidal maniac?

In one simple Post, Debe Terhar shows herself to:

  • Have a serious lack of professionalism
  • Questionable judgment
  • The inability to apologize
  • Lacking in historical understanding
  • A belief that we are stupid enough to believe she wasn't comparing the President to Hitler

Apparently in Terhar's world, only educators are to be held accountable.

The Science of Value-Added Evaluation

"A value-added analysis constitutes a series of personal, high-stakes experiments conducted under extremely uncontrolled conditions".

If drug experiments were conduted like VAM we might all have 3 legs or worse

Value-added teacher evaluation has been extensively criticized and strongly defended, but less frequently examined from a dispassionate scientific perspective. Among the value-added movement's most fervent advocates is a respected scientific school of thought that believes reliable causal conclusions can be teased out of huge data sets by economists or statisticians using sophisticated statistical models that control for extraneous factors.

Another scientific school of thought, especially prevalent in medical research, holds that the most reliable method for arriving at defensible causal conclusions involves conducting randomized controlled trials, or RCTs, in which (a) individuals are premeasured on an outcome, (b) randomly assigned to receive different treatments, and (c) measured again to ascertain if changes in the outcome differed based upon the treatments received.

The purpose of this brief essay is not to argue the pros and cons of the two approaches, but to frame value-added teacher evaluation from the latter, experimental perspective. For conceptually, what else is an evaluation of perhaps 500 4th grade teachers in a moderate-size urban school district but 500 high-stakes individual experiments? Are not students premeasured, assigned to receive a particular intervention (the teacher), and measured again to see which teachers were the more (or less) efficacious?

Granted, a number of structural differences exist between a medical randomized controlled trial and a districtwide value-added teacher evaluation. Medical trials normally employ only one intervention instead of 500, but the basic logic is the same. Each medical RCT is also privy to its own comparison group, while individual teachers share a common one (consisting of the entire district's average 4th grade results).

From a methodological perspective, however, both medical and teacher-evaluation trials are designed to generate causal conclusions: namely, that the intervention was statistically superior to the comparison group, statistically inferior, or just the same. But a degree in statistics shouldn't be required to recognize that an individual medical experiment is designed to produce a more defensible causal conclusion than the collected assortment of 500 teacher-evaluation experiments.

How? Let us count the ways:

  • Random assignment is considered the gold standard in medical research because it helps to ensure that the participants in different experimental groups are initially equivalent and therefore have the same propensity to change relative to a specified variable. In controlled clinical trials, the process involves a rigidly prescribed computerized procedure whereby every participant is afforded an equal chance of receiving any given treatment. Public school students cannot be randomly assigned to teachers between schools for logistical reasons and are seldom if ever truly randomly assigned within schools because of (a) individual parent requests for a given teacher; (b) professional judgments regarding which teachers might benefit certain types of students; (c) grouping of classrooms by ability level; and (d) other, often unknown, possibly idiosyncratic reasons. Suffice it to say that no medical trial would ever be published in any reputable journal (or reputable newspaper) which assigned its patients in the haphazard manner in which students are assigned to teachers at the beginning of a school year.
  • Medical experiments are designed to purposefully minimize the occurrence of extraneous events that might potentially influence changes on the outcome variable. (In drug trials, for example, it is customary to ensure that only the experimental drug is received by the intervention group, only the placebo is received by the comparison group, and no auxiliary treatments are received by either.) However, no comparable procedural control is attempted in a value-added teacher-evaluation experiment (either for the current year or for prior student performance) so any student assigned to any teacher can receive auxiliary tutoring, be helped at home, team-taught, or subjected to any number of naturally occurring positive or disruptive learning experiences.
  • When medical trials are reported in the scientific literature, their statistical analysis involves only the patients assigned to an intervention and its comparison group (which could quite conceivably constitute a comparison between two groups of 30 individuals). This means that statistical significance is computed to facilitate a single causal conclusion based upon a total of 60 observations. The statistical analyses reported for a teacher evaluation, on the other hand, would be reported in terms of all 500 combined experiments, which in this example would constitute a total of 15,000 observations (or 30 students times 500 teachers). The 500 causal conclusions published in the newspaper (or on a school district website), on the other hand, are based upon separate contrasts of 500 "treatment groups" (each composed of changes in outcomes for a single teacher's 30 students) versus essentially the same "comparison group."
  • Explicit guidelines exist for the reporting of medical experiments, such as the (a) specification of how many observations were lost between the beginning and the end of the experiment (which is seldom done in value-added experiments, but would entail reporting student transfers, dropouts, missing test data, scoring errors, improperly marked test sheets, clerical errors resulting in incorrect class lists, and so forth for each teacher); and (b) whether statistical significance was obtained—which is impractical for each teacher in a value-added experiment since the reporting of so many individual results would violate multiple statistical principles.

[readon2 url="http://www.edweek.org/ew/articles/2013/01/16/17bausell.h32.html"]Continue reading...[/readon2]

$50 million. 3 years. No clue.

More on that awful Gates study

Though science does sometimes prove things that are not intuitive, science does depend on accurate premises. So, in this case, IF the conclusion is that “you can’t believe your eyes” in teacher evaluation — just because you watch a teacher doing a great job, this could be a mirage since that teacher doesn’t necessarily get the same ‘gains’ as the other teacher that you thought was terrible based on your observation — well, it could also mean that one of the initial premises was incorrect. To me, the initial premise that has caused this counter-intuitive conclusion is that value-added — which says that teacher quality can be determined by comparing student test scores to what a computer would predict those same students would have gotten with an ‘average’ teacher — is the faulty premise. Would we accept it if a new computer programmed to evaluate music told us that The Beatles’ ‘Yesterday’ is a bad song?

One thing that struck me right away with this report is that the inclusion of student surveys — something that aren’t realistically ever going to be a significant part of high stakes teacher evaluations — is given such a large percentage in each of the three main weightings they consider (these three scenarios are, for test scores-classroom observations-student surveys, 50-25-25, 33-33-33, and 25-50-25.)

Conspicuously missing from the various weighting schemes they compare is one with 100% classroom observations. As this is what many districts currently do and since this report is supposed to guide those who are designing new systems, wouldn’t it be scientifically necessary to include the existing system as the ‘control’ group? As implementing a change is a costly and difficult process, shouldn’t we know what we could expect to gain over the already existing system?

[readon2 url="http://garyrubinstein.teachforus.org/2013/01/13/50-million-3-years-no-clue/"]Read the whole piece[/readon2]