SATs Scored in Error by Test Companies Roil Admissions Process
A PDF file contains several substories along with this major thread.
By David Glovin and David Evans
Nov. 3 (Bloomberg) -- Jerry Lee Faine Elementary School in Dothan, Alabama, starts each day with two hours of reading and vocabulary. After that, there's arithmetic. ``If you can read, you can do anything,'' says Principal Deloris Potter, a spry woman of 59 who has run the school since 2002.
Potter, trusting the work of her teachers, was confident of passing grades in April 2005 as students began two weeks of mandatory standardized testing in reading and math. That July, state education officials told Potter her school had failed the Alabama Reading and Mathematics Test. The state warned it might fire teachers if scores didn't improve, she says. A dozen students transferred after the substandard rating. Faculty morale plunged.
``We felt like dogs,'' says Charlotte Adams, a reading specialist at the school.
In February 2006, the state said Jerry Lee Faine Elementary had passed. Harcourt Assessment Inc., a unit of London-based Reed Elsevier Plc and one of the world's largest test companies, had improperly graded the exam.
The snafu is at least the 30th time since 2000 that San Antonio-based Harcourt Assessment, which also wrote the exam, has made errors such as improper scoring, faulty instructions and questions with more than one answer.
Harcourt isn't alone. Other companies are constructing flawed tests, administering them improperly and scoring them incorrectly, according to lawsuits and education department records in 15 states.
In March, Pearson Assessments, a unit of London-based Pearson Plc, the world's biggest educational publisher, had to explain to high schoolers across the U.S. that it had erred in scoring about 5,000 SAT college entrance exams because its scanners couldn't read answer sheets that had expanded from humidity.
The next month, education officials in Minnesota discovered a separate issue with answer sheets that Pearson Assessments had created for a state-mandated exam. At least 500,000 people taking tests from 2000 through 2006 -- from Nevada third graders to aspiring teachers in many states -- were victims of test company mistakes, documents show.
``The errors we've seen from testing companies are probably just the tip of the iceberg,'' says David Berliner, 68, Regents' Professor of Education at Arizona State University in Tempe, who has written more than 200 articles, books and book chapters about education and served as president of the 25,000-member American Educational Research Association. ``State education departments often lack the ability to adequately supervise these companies.''
Millions of Tests
The U.S. is in a testing frenzy. Students in the 92,816 American public schools will take at least 45 million standardized reading and math exams this year. That will jump to 56 million in the 2007-08 school year, when states begin testing science as part of the 2002 federal No Child Left Behind law, the most comprehensive education overhaul in half a century.
Beyond No Child, tens of millions of additional tests assess college hopefuls, certify future stockbrokers and even evaluate preschoolers. With the stakes for making the grade so high for so many, errors by test companies have dramatic consequences.
Joseph Conigliaro lost his Pennsylvania teaching job after Princeton, New Jersey-based Educational Testing Service, the world's biggest standardized test company, incorrectly scored three of his licensing exams. ETS, which will pay $11.1 million to 4,100 teachers who were falsely failed, called the error an ``anomaly.''
Ryan Beck & Co. asked Linda Cutler to resign from a senior associate job at the securities firm after she and 1,881 other test takers were scored incorrectly last year on the Series 7 licensing exam for securities representatives.
``It's an exponentially growing catastrophe,'' says James Popham, an emeritus professor of education at University of California, Los Angeles, and author of 25 books on education. ``No one knows how bad it is, and it's going to get worse.''
Deputy U.S. Education Secretary Raymond Simon says states must better oversee test companies. ``The whole teaching system is based on the results of those tests,'' Simon, 61, says. ``If the integrity of the testing process is called into question, that brings into question the whole accountability system.''
The national obsession with performance and measurement means a booming business for test-producing and grading companies. In 2005, CTB/McGraw-Hill, Educational Testing Service, Harcourt Assessment, Pearson Assessments and smaller firms generated $2.8 billion in revenue from testing and test preparation, according to Boston-based research firm Eduventures LLC. No Child tests alone produced about $500 million in annual revenue in 2005-06.
Along with creating exams, Harcourt Assessment, Pearson Assessments and companies such as White Plains, New York-based Haights Cross Communications Inc. sell mass-produced workbooks, practice tests and computer software that teachers use year-round to prepare students for No Child and other tests.
The burgeoning test preparation industry generated $1.7 billion in annual revenue last year. The $1.1 billion testing market and the $1.7 billion test prep business will grow by a combined 30 percent by the 2009-10 school year, Eduventures predicts.
For test companies, pitching schools to buy preparation materials after receiving a No Child contract is routine, says Robert Schaeffer, public education director at the National Center for Fair & Open Testing, a Cambridge, Massachusetts-based non-profit group.
``It's standard business practice, the equivalent of razor companies' giving away razors so they can make money selling blades,'' he says. ``It's where the real profits are.''
Profit margins in test preparation are as much as seven times higher than they are for No Child tests, partly because there are no requirements for high-quality questions on practice exams. States leave it to schools and school districts to decide whether the test preparation materials they're buying are sound.
Haights Cross, publisher of the Buckle Down test preparation workbooks, reported operating margins of 21 percent in its test preparation division for the first half of 2006.
In comparison, No Child tests, which must be custom designed for almost every state, have pretax profit margins as low as 3 percent, says Kurt Landgraf, chief executive officer of Educational Testing Service. He says his not-for-profit company lost $2.6 million on a $236 million four-year No Child contract in California.
Richard Rizzo, chief financial officer of Measured Progress Inc., a Dover, New Hampshire-based nonprofit firm that produces No Child tests, says he expects to earn margins triple those of No Child exams by selling practice questions and tests that schools use to gear up for the actual exams. Getting a foot in the door with a No Child contract can also lead to sales of achievement or psychological tests not related to No Child.
``Companies could conceivably low-ball the customized test because they know they could go in and sell the off-the-shelf products with a 40-50 percent margin,'' says Rizzo, 62, referring to tests that aren't specially designed for individual states.
Whether or not they low-ball, companies often scrimp when they bid on No Child contracts, Eduventures analyst Tim Wiley says. Getting a contract involves the same process as selling supplies or cafeteria food to a school: A company submits what it expects to be a winning package. ``As with any bidding situation, it definitely requires a lot of cost cutting,'' Wiley says. ``Or, in some cases, cutting corners.''
In Florida, CTB/McGraw-Hill won part of the state's testing contract for 3,800 schools in 2005. To grade the essay portion, the Monterey, California-based unit of McGraw-Hill Cos. hired $10-an-hour workers from Kelly Services Inc., the second-largest U.S. provider of temporary employees, and other companies.
Among the 2,947 graders was a person who won the job while he was employed packing bags of potato chips for PepsiCo Inc.'s Frito-Lay unit, applications compiled by the Florida state senate show. Kelly spokeswoman Renee Walker declined to comment.
Another grader was a cook in an Orlando, Florida, diner. One essay evaluator wrote he was ``layed off'' from a clerical job after working as a janitor. He graduated from Ambassador University, a Worldwide Church of God-run school in Big Sandy, Texas, in 1997. The school shut down that same year. Another said that he majored in ``Phylosophy/Humanity'' at Mount Angel Seminary in St. Benedict, Oregon.
Steven Weiss, vice president for communications at McGraw- Hill, said in an e-mailed statement that the company had performed extremely well in scoring more than 90 million documents with a total of more than 755 million essay and short- answer questions during the past five years.
CTB/McGraw-Hill, Harcourt Assessment and Pearson Assessments don't break down their revenue from No Child tests and preparation materials in regulatory filings. Public records from the Wyoming department of education show the state is paying Harcourt, which has a $21 million, four-year No Child contract, more than $120 per student each year. Of that, about half is for No Child tests, and the rest is for preparation materials and other testing products.
Schools in Okaloosa County, Florida, pay $9.50 per student for a series of preparation tests called Stanford Learning First, which Harcourt Assessment renamed Learnia. By comparison, Harcourt received $4.93 per child from the state of Florida in 2005 to develop questions for its No Child-mandated annual Comprehensive Assessment Test.
Harcourt Assessment's experience shows how winning a No Child bid can be a prelude to more sales. In 2004, Harcourt got a four-year, $44.5 million contract to develop and score Illinois's No Child exams. Chicago schools then began purchasing Harcourt materials, testing director Xavier Botana says.
The preparation products included Stanford Learning First practice tests that measured student progress as they prepared for No Child exams. In the 2005-06 school year, the district spent $1.8 million on Harcourt's new Stanford Learning First product.
Christine Rowland, a former teacher of English as a second language who now trains colleagues at Christopher Columbus High School in the Bronx, New York, says her pupils didn't learn more because of increased testing. Still, she relied on test preparation materials to help students pass the math test. The cost of failure was too high, she says.
``If I know they are going to test six things six weeks from now, that's what I'm going to teach,'' Rowland, 46, says. ``It puts a tremendous amount of pressure on. The real fear is that it turns students off from learning.''
Test companies, aware that Rowland and other teachers are being judged by how students do on No Child exams, are inundating schools with ads for preparation products such as practice tests, software and banks of sample questions. Often they say their materials are designed specifically to help students pass the state's No Child test.
``I'm getting mail from companies I've never heard of,'' says Susan Friedwald, head of teacher training at Public School 48 in the Bronx.
At Cracker Trail Elementary School in Sebring, Florida, 11- year-old Alexis Szoka took dozens of practice exams last year leading up to the Florida Comprehensive Assessment Test. She wound up a nervous wreck. ``My daughter has such test anxiety, she can't take a test anymore,'' says Alexis's mother, Carol Szoka.
One exam measured whether Alexis understood vocabulary and another checked her spelling. The school tested how well she read and whether she knew math. Some tests compared her reading and math skills with those of other fourth graders. Alexis was evaluated on phonics, writing and her understanding of text on a computer. Most tests were given two, three or four times a year. Teachers gave chapter tests in reading and math and benchmark tests throughout the year to see whether Alexis was progressing.
Andrew Lethbridge, Cracker Trail's vice principal, says one test gave fourth graders practice in filling in answer sheet bubbles on other tests. The materials came from divisions of Harcourt Assessment, Pearson Assessments and smaller, privately held companies.
``It was never like this,'' says Carol Szoka, who has two grown children who went through the same schools in Sebring, which is 85 miles (137 kilometers) south of Orlando. ``They had an achievement test. They just took it. They weren't prepped.''
Richard Demeri, Cracker Trail's principal, says test preparation materials have helped his students. Seven years ago, the school was given a grade of C by the state. Now, with test scores higher, the school has an A from the state and is no longer on probation.
``There's very little spray-and-pray teaching going on -- where you spray everybody and pray they get it,'' he says. The school uses test results to analyze each student's progress. ``It's much more individualized now,'' he says.
Even if Demeri's students are prepared to take No Child tests, two Florida state senators question whether CTB/McGraw- Hill has qualified people to grade them.
Senators Walter ``Skip'' Campbell and Leslie ``Les'' Miller Jr. sued the state education department and CTB/McGraw-Hill earlier this year to obtain applications of test graders. The department had refused to release the applications, citing confidentiality. CTB/McGraw-Hill settled the suit by providing copies of the scorers' personnel files with personal identifying information removed.
CTB/McGraw-Hill's $82 million, three-year Florida contract requires a scorer to have a bachelor's degree in mathematics, reading, science, education or a related field. On its Web site, the Florida Department of Education assures parents that graders of the Florida Comprehensive Assessment Test are professional, trained scorers.
Information the senate obtained shows one grader had an associate's degree, which is below a bachelor's, from the University of Delhi's School of Correspondence Courses and Continuing Education in Delhi, India.
She worked as a $7.50-an-hour cashier at a duty-free shop at O'Hare International Airport in Chicago before being hired to grade exams, according to the settlement documents. CTB/McGraw- Hill now says this person never scored exams.
A personal trainer with a degree in sports science from the University of Leipzig in Germany also graded essays, as did a convicted shoplifter who graduated from West Virginia University with a degree in physical education, the applications show. A person from Hungary wrote he was a ``pyshical education'' major. A physical education major from Methodist College in Fayetteville, North Carolina, wrote that she had attended ``Methidist College.''
McGraw-Hill's Weiss said its scorers from the University of Delhi met the requirements for a bachelor's degree. ``Individuals must undergo a comprehensive training process before becoming qualified to score,'' Weiss wrote. ``Scorers must maintain performance quality throughout the process.''
Spelling Not a Requirement
CTB/McGraw-Hill spokeswoman Kelley Carpenter says the company subjects scorers to a rigorous three- to five-day training program. Next year, at Florida's request, the company will ensure that scorers have appropriate backgrounds for the subjects they grade, she says.
``They are constantly monitored,'' she says. ``And if they don't match the quality performance standards, they're not retained as scorers.''
Carpenter says spelling errors on an application don't disqualify someone from being hired as a scorer. ``Spelling in and of itself is not a requirement,'' she says.
When Deputy Education Secretary Simon is shown misspellings on applications of Florida scorers, he says he would demand excellence. ``It's absolutely important that the integrity of the scorers is something the companies would be proud of and feel comfortable with,'' he says. ``I can't imagine they would feel comfortable with a non-speller.''
Cornelia Orr, head of the Florida Office of Assessment and Performance, says she reviewed about 25 percent of the grader applications. ``I felt like CTB had minimally met our expectations,'' she says. ``There are ways they can improve.''
One reason for the testing foul-ups and their dire effects is that there's no federal oversight of the testing industry. When the U.S. Congress authorized the No Child law it didn't create an agency to evaluate whether the companies making and selling the exams do an adequate job. Each state oversees its own test contractor.
Roderick Paige, who ran the No Child program as U.S. education secretary from 2001 to 2004, says the law is a good one. He says his concern is that testing may not be done accurately and competently. Paige, 73, says he summoned top executives from 20 testing companies to a conference room at the U.S. Department of Education on Feb. 20, 2003, and demanded better performance.
In 2005, the Education Department's inspector general announced plans to study whether there's a need for federal review to detect and prevent errors. The study isn't yet under way, spokeswoman Catherine Grant says.
``We've got to get better testing producers,'' says Paige, who's now chairman of Chartwell Education Group LLC, a Washington-based school consulting company. ``They're making mistakes.''
Harcourt Assessment is making the most errors, according to records in 15 state education departments. In addition to erroneously failing Jerry Lee Faine Elementary, Harcourt wrongly flunked three other Alabama schools because of its grading snafu. It mistakenly passed 10 Alabama schools that should have failed, the state said.
In Connecticut, Harcourt Assessment reported the wrong reading test scores for 355 high school students in 51 districts last year. The state fined the company $80,000. In Georgia, Hawaii, Illinois, Massachusetts and Virginia, Harcourt made errors on No Child tests and achievement tests given to measure how students compared with one another. States fined the company hundreds of thousands of dollars.
``Employees took shortcuts,'' Harcourt Assessment Senior Vice President Robin Gunn wrote in a May 28, 2004, letter to Hawaii school principals, promising stricter oversight. Gunn has since left the company.
Hawaii hired a not-for-profit firm, Washington-based American Institutes for Research, to develop and score the tests after discovering more errors on Harcourt's 2005 exams. Illinois also replaced Harcourt in the middle of its contract; Connecticut, Massachusetts and Virginia didn't renew their contracts with the company.
Nevada fired Harcourt in 2004, after the company mistakenly failed hundreds of students, gave inflated scores to thousands of others and produced tests with missing pages, misspellings and flawed instructions, according to Nevada Education Department records.
Fined and Fired
``It was errors, one after the other, and not to a single student but to a large number,'' says Karlene Lee, the assistant superintendent in Clark County, Nevada, which includes Las Vegas. ``In education, we don't have the luxury to say that 2 percent doesn't matter. Every child has to be accurate.''
Nevada fined Harcourt Assessment $425,000 in 2002, before firing the company.
Harcourt's approximately $290 million in revenue last year was 3 percent of Reed Elsevier's sales, according to company filings. Reed Elsevier reported its profit increased 62 percent in the six months ended on June 30 to 217 million pounds ($403 million) compared with a year earlier. The company's shares rose 7.7 percent this year to 588 pence on Nov. 2.
Harcourt Assessment hired a new CEO, Michael Hansen, who took over in July after serving as executive vice president for corporate development at Guetersloh, Germany-based Bertelsmann AG, Europe's largest media company. Hansen, 45, says his company won't slip up again. He blames errors on the enormous demand for made-to-order state tests.
``You went from an industry that was largely standardized to an industry that was highly, highly customized,'' Hansen says during an interview in a conference room in his San Antonio office suite, which is adjacent to the test production work floor. ``Our most sacred obligation is that the test results are accurate and that they are timely.''
Last year privately held Measurement Inc., a Durham, North Carolina-based test development and scoring company, wrongly failed 890 students out of the 5,461 it tested on Ohio's high school graduation exams. The company says it scored the exams correctly and then erred when it determined the students' grades based on the number of questions they answered correctly.
``We had a really spotless reputation,'' Senior Vice President Mike Bunch says. ``This was just devastating to us.''
Pearson Assessments grades 40 million exams each year. The company has the high-profile job of scoring the SAT, which more than 3,000 colleges and universities use as a gauge for admitting students.
`Hard to Get Perfection'
Pearson discovered its SAT scoring error in January after two students asked that their results be hand-scored. Score changes affected about 1 percent of the October 2005 test takers, says the New York-based College Board, a nonprofit group that represents 5,000 colleges and oversees the exam. Before most college admission decisions were announced, the College Board re-reported the roughly 4,400 scores that had been underscored.
``When you do 12 million tests a year, a lot of people are involved in that,'' College Board President Gaston Caperton says. ``It's very hard to get perfection.''
Shane Fulton, a lean youth who played soccer and tennis at George School in Newtown, Pennsylvania, knows the pain of an incorrect score. Fulton had his sights on attending New York University or Lafayette College in Easton, Pennsylvania.
`Something Was Wrong'
In June 2005, at the end of his junior year at the Quaker- run high school, he took his first SAT. He earned a score of 1,910 out of 2,400 on the three-part test, which assesses mathematics, reading and writing. Not satisfied with his performance on the math portion, he took the test a second time in October. He was shocked when the grade came back as a 1,330.
``I knew that something was wrong,'' says Fulton, 19, of Yardley, Pennsylvania. He asked to have his exam graded by hand. When the results were returned more than a month later, his score was actually a 1,720, or 390 points higher than initially reported.
By then, Fulton had suffered restless nights, sought sleeping pills from his parents and broken down in tears because of the uncertainty surrounding the scores and his future. Adding to his anxiety, he'd taken the SAT a third time because he didn't yet know his results on the second test. On that one, he earned an 1,850.
``Every year, there's more of an emphasis on how you do,'' says Fulton, who's attending Northeastern University in Boston and is suing Pearson Assessments and the College Board over the error. ``I was thinking I wouldn't get into any of the colleges I applied to.''
$50 Million Investment
Mistakes may soon cost Pearson Assessments and other test companies business. Educational Testing Service wants to bring scoring in-house to reduce the chance of errors. ETS's Landgraf has directed the company to invest $50 million so it can expand its scoring operation within three years. He estimates that will produce $33 million in new annual revenue. Pearson shares gained 12 percent this year as of Nov. 2, to 769.5 pence.
Having ETS grade his exam didn't help Pennsylvania teacher Conigliaro, one of the 4,100 false failures on the Praxis test. Forty-four states require the Praxis to evaluate teaching skill and knowledge in a particular field. ETS developed the Praxis and then, in Conigliaro's case, scored it incorrectly -- multiple times.
Conigliaro, 55, an engineer and former machine shop owner, started teaching seven years ago as an intern at Mountain View Junior/Senior High School in Kingsley, Pennsylvania. His employment there was contingent on his passing the Praxis to get final certification. He took the exam in April 2003 and was told he'd failed. He took it again and got a second failing score. He took it a third and a fourth time and again flunked.
``I was missing by one or two points each time,'' he says.
Conigliaro was fired from his teaching job and wound up working as a bartender. ``I didn't want to leave the house for a year and a half because I was too embarrassed,'' he says.
ETS notified Conigliaro in July 2004 that there were scoring errors on his tests and that he had actually passed. In a press release that month the company cited a ``statistical anomaly'' in the scoring of nine exams from January 2003 to April 2004 and apologized to test takers. ETS spokesman Tom Ewing declined to comment further.
According to court papers by teachers who later sued ETS in federal court in New Orleans, the firm didn't start an investigation of its scoring of short essays until an unnamed state challenged the results. In March, the company agreed to pay $11.1 million to the test takers to settle the lawsuit.
Conigliaro, who sued and was part of the settlement, says he would have succeeded on at least three of the four exams he was told he'd failed. ``Yes, I'm bitter,'' says Conigliaro, who, after passing the Praxis and getting his license, now teaches business and accounting at Blue Ridge High School in New Milford, Pennsylvania. ``I was just about to get tenure, and I had to start all over again.''
Errors can occur in the earliest stages of the test-making process and then snowball. In 2003, the Minnesota Department of Education found flaws in questions proposed by Maple Grove, Minnesota-based Data Recognition Corp., a privately held firm that provides testing for eight states. Minnesota school officials reviewed some questions, which are known as items. About 6 percent had no correct answers or multiple correct answers.
``There are other concerns about item quality with another 60-70 percent,'' testing director Reginald Allen wrote in 2003 after the company challenged the state's decision not to renew its contract. The flawed test questions didn't make it onto state exams.
Company lawyer Dwight Rabuse declined to comment except to say that the state later hired a Data Recognition staffer to replace Allen. Minnesota Education Department spokesman Randy Wanke declined to comment. Minnesota now contracts with Pearson Assessments to provide its state tests.
``When you have an education reform agenda that's relying so heavily on standard tests to ensure school quality, it doesn't take so many problems to undermine credibility,'' says Thomas Toch, co-director of Washington-based research firm Education Sector, who wrote a 2006 report on test errors.
Executives at testing companies say they strive for perfection in the face of state demands for new tests each year, in at least two different subjects and for seven different grades.
Potential for Errors
Stuart Kahl, president and founder of Measured Progress, says the industry uses dozens of quality checks as companies draft, edit, print and deliver exams; retrieve, scan and read papers; and calculate, compare and convert raw scores into test grades. The process may take two years from start to finish.
``There's no question there are tremendous demands placed on the industry,'' Kahl says. ``Obviously, when you redo things every year, you have tremendous potential for errors.''
Former Harcourt Assessment President Jeff Galt says state education departments are sometimes to blame for errors that they require their testing contractors to assume responsibility for. He points to Connecticut, which is using Measurement Inc., its third testing contractor since 2003. The state got rid of Harcourt and then parted ways with CTB/McGraw-Hill.
``You have to wonder, Is the problem with the testing company or with the department?'' says Galt, 50, who now teaches business at the University of the Incarnate Word in San Antonio. Connecticut Education Department spokesman Henry Garcia declined to comment.
Harcourt Assessment's inability to follow instructions from Alabama is what cost Jerry Lee Faine Elementary its good name.
After the school was notified of its failure to make the required adequate yearly progress, the state placed it in the category of School Improvement, as probation is called under the No Child program. Newspapers publicized the designation, and parents won permission to transfer children to other schools.
``People will not move into this community,'' says Alfreda Mays-Rogers, whose grandchild is in first grade at the school.
The Alabama Department of Education summoned principal Potter 100 miles north to Montgomery, she says. Officials demanded more teacher training and insisted on additional reading instruction. Potter says she researched curricula used by other schools and dissected years of test data to figure out why her pupils hadn't passed. Nothing stood out.
During Potter's crisis of confidence, Kirby Hubbard, the testing director in Etowah County, about 250 miles to the north, discovered that Harcourt Assessment had miscalculated his schools' No Child results.
`Mad, Mad, Mad'
Harcourt had tallied the scores of students who'd been absent during part of the exam week, failing to follow Alabama's instruction to count the scores of only students who took the entire multipart test, state Education Superintendent Joseph Morton said in a Nov. 8, 2005, letter to Harcourt. That same type of error affected Jerry Lee Faine Elementary.
When the state told Potter her school had actually passed on Feb. 9, 2006, she took to the school intercom and made the announcement. Teachers ran into the hallways, cheering.
``We were happy, happy, happy,'' Potter says. ``But you turn to the other side, we were mad, mad, mad.''
Along with Potter, educators in Florida, Nevada and across the U.S. have to live with test company mistakes every year. Boston College emeritus professor George Madaus and researcher Kathleen Rhoades say there should be independent oversight of crucial exams.
``There's so much error in these products,'' Rhoades says.
`Make Testing Better'
Madaus, co-author of a 2003 study on test errors, envisions an impartial federally financed panel that would monitor state testing programs to ensure they're well crafted and used correctly. Such a board would analyze why there are errors and how they can be minimized. It also may offer a seal of approval on the test preparation products flooding the market, which can generate such a big chunk of a test company's earnings.
``This is not anti-testing,'' Madaus says. ``This is an attempt to make testing better.''
Potter tries not to be bitter. She notes with pride how her school has now passed the state test for two consecutive years. She has a message for test companies. ``They're hurting students more than anything else,'' she says. ``Please don't make that mistake on students. That's a reflection on our school, on my students, on my teachers. That's a reflection on me.''
It's also a reflection on the $2.8 billion test industry, which profits from selling materials to prepare students for high-stakes exams it has a hard time getting right.
To contact the reporters on this story: To contact the reporters for this story: David Evans in Los Angeles email@example.com David Glovin in New York at firstname.lastname@example.org
David Glovin and David Evans
INDEX OF NCLB OUTRAGES