| 
				
                
                 
                 
                The Plain 
				Dealer, August 24, 2008  
				 
				Tests measure what students know. Like a Polaroid, they give a 
				snapshot of knowledge frozen at one moment in time. 
				
                 
				But what if you could measure how much a child learns over the 
				course of a school year? What if you could gauge what a school 
				actually adds to a child's learning experience?  
				
                 
				In Ohio, you can.  
				
                 
				This year's district and school report cards, which will be 
				released Tuesday by the Ohio Department of Education, for the 
				first time will include a measurement known as value-added.  
				
                 
				The revolutionary formula, designed more than two decades ago by 
				a homespun statistical guru from the rolling hills of eastern 
				Tennessee, has rocked the education world.  
				
                 
				Put simply, value-added tracks whether a year's worth of 
				learning is actually happening in the course of a school year -- 
				regardless of whether a child passes a test at the end of that 
				year. 
				
                 
				Ohio schools and districts that exceed expected growth will be 
				rewarded with a plus sign on their report cards and could move 
				up a rung on the ladder of academic ranking. Those who make 
				expected growth will get a check mark. Those who miss their 
				target will get a minus sign.  
				
                 
				More important, the formula will show parents which schools are 
				really teaching children something and which schools are 
				coasting.  
				
                 
				It will reveal that in some affluent districts that easily hit 
				state testing targets, students aren't being challenged enough.  
				
                 
				And it will disclose that students in some high-poverty 
				districts are making amazing gains, even if they aren't yet 
				passing state tests.  
				
                 
				Think of it this way: Many a parent has used pencil marks on a 
				bedroom wall to plot a child's growth. Value-added does the same 
				thing, measuring academic growth rather than penalizing a child 
				for not making a predetermined height like, say, 6-foot-2.  
				
                 
				"It is, at this point in time, the most robust methodology that 
				has been developed," said William Sanders, who began working on 
				the concept 25 years ago as a University of Tennessee 
				researcher. "It's fair to hold adults accountable for the 
				progress rate of children."  
				
                 
				Ohio and just three other states -- Tennessee, Pennsylvania and 
				North Carolina -- use the value-added measurement statewide. 
				Ohio's valued-added data is based on the reading and math scores 
				of public-school students in grades 4-8.  
				
                 
				A Columbus-based nonprofit, Battelle for Kids, began the 
				value-added experiment in Ohio in 2002 with a pilot project 
				using the concept in 42 volunteer school districts. By last 
				year, the experiment had evolved into the largest value-added 
				pilot program in the United States.  
				
                 
				State officials hired Sanders -- an affable professor with an 
				easy drawl and a keen mind for numbers -- as a consultant. 
				Together they fashioned a formula that fit Ohio's system.  
				
                 
				What makes Sanders' formula compelling -- and controversial -- 
				is its ability to predict what a student should achieve. When 
				actuaries compute life expectancy, they take into account 
				factors such as a person's family history, or habits such as 
				smoking or hobbies such as skydiving. Likewise, value-added 
				takes into account a student's academic background and comes up 
				with a predicted score.  
				
                 
				"This is the most important thing we do, in my view," Sanders 
				said. "I'm not going to get them all right, but if you bet 
				against me, you're going to lose."  
				
                 
				It also represents a seismic shift in education research. For 
				decades, educators embraced the thinking of the late sociologist 
				James Coleman, who argued that a child's socioeconomic status 
				largely determined academic achievement and that what schools do 
				doesn't matter much.  
				
                 
				Sanders turned that thinking on its head. Instead of adjusting 
				for the income level of a child's family, Sanders compares the 
				child against his or her own past academic performance.  
				
                 
				Because all children don't achieve at the same level, schools 
				are evaluated on what "value" they add to a student's 
				experience. Schools that don't add anything need to ask why.  
				
                 
				"If a district is below the expected level of growth, it needs 
				to take account of what's going on," said Matt Cohen, executive 
				director for policy and accountability for the Ohio Department 
				of Education.  
				
                 
				Not everyone is completely enamored of Sanders' work. Some, such 
				as Virginia-based education researcher Gerald Bracey, contend 
				value-added is flawed because it is still based on test scores. 
				And the Education Trust's Kati Haycock warns that while growth 
				is important, it's also crucial that all children -- poor, rich, 
				black, white -- eventually master academic skills.  
				
                 
				"It's not just growth," said Eric Gordon, chief academic officer 
				of the Cleveland public schools. "It's growth and standards."  
				
                 
				But the most potentially explosive thing about value-added is 
				its ability to determine which individual teachers are effective 
				and which are not. Some school districts, such as Houston, are 
				already using Sanders' work as the foundation for a performance 
				pay system for teachers.  
				
                 
				In Ohio, Battelle for Kids is using valued-added in a volunteer 
				pilot program involving about 1,800 teachers in about 240 
				schools across the state. What it has found: effective teaching 
				is not limited to the state's affluent districts with the 
				highest test scores.  
				
                 
				The Ohio pilot is not being used to evaluate or pay teachers but 
				rather to learn what works in the classroom and what doesn't, 
				said Jim Mahoney, executive director of Battelle for Kids.  
				
                 
				"The worst thing that can happen is that you're given this tool 
				and you make it into a weapon," Mahoney said. "The fact that a 
				tool might be misused is not a good reason to do away with it."
				 
				 
				Questions and answers on value-added measurement of students 
				These are some frequently asked questions about the value-added 
				measurement: 
				 
				Q. How is value-added different from traditional measures of 
				student performance? 
				Student performance can be measured in two very different ways, 
				both of which are important. Achievement describes the absolute 
				levels attained by students in their end-of-year tests. Growth, 
				in contrast, describes the progress in test scores made over the 
				school year. 
				 
				Q/ Why is measuring both achievement and progress important? 
				In the past, students and schools have been ranked solely on 
				achievement. The problem with this method is that achievement is 
				highly linked to the socioeconomic status of a student's family. 
				In contrast, value-added assessment measures growth and answers 
				the question: How much value did the school staff add for 
				students?  
				 
				Q. Is it possible to show progress with all students -- special-ed, 
				gifted and low-performing 
				Yes. The value-added methodology used is sensitive to individual 
				students' achievement levels. It measures growth from the end of 
				one year to the end of the next year, regardless of whether a 
				student performs below or above grade level. 
				 
				Q.Does value-added analysis require additional testing?  
				No. Value-added analysis uses existing standardized test data 
				and can be done only where annual testing is provided. 
				 
				Q. Do socioeconomic factors affect valued-added measurements? 
				No. Leading experts have shown that those factors -- such as 
				household income or parents' education level -- have no 
				significant relationship with student progress measures. 
				 
				SOURCE: Battelle for Kids; University of Pennsylvania.  
                
                  
                
                                
                
                back to the top     ~    
                back to Breaking News     
                ~     back to 
                What's New 
                   |