Critical Issues in Evaluating the Effectiveness of Technology

By Mary McNabb, Mark Hawkes, and Ullik Rouk

 

Reaction Paper

By Sandy Nightingale

CET720

 

     At the Secretary’s Conference on Educational Technology on July 12-13, 1999 in Washington, D.C., evaluating the effectiveness of technology was a topic of discussion.  The article stated that in recent years there has been a change in schools’ focus on technology, shifting away from building and implementing a technology infrastructure towards evaluating the effectiveness of its use in schools and classrooms.  It also went on to say that if funding is to continue, the effectiveness of technology’s use in schools and classrooms must be evaluated.  I agree that we not only need summative evaluations for the state policymakers, but also a more useful formative evaluation for school practitioners.  My school district seems to be stuck between the two phases of implementing technology and evaluating its effectiveness. 

     Seven critical issues were identified at the conference, and I will react to each issue and compare them to the technology processes currently being used in my district, although I must emphasize again that our evaluation process is at somewhat of a standstill.

 

Critical Issue 1:  “The effectiveness of technology is embedded in the effectiveness of other school improvement efforts.”

     As the article stated, “Developing ways to isolate the effects of technology within a dynamic environment where so many elements work together is one of evaluation’s most challenging issues.”  I believe that it is very difficult to produce measurable results that portray the effects of technology.  I agree that evaluation designs need to take the “localness” of evaluation into account.

     My school district’s Technology Plan (White Lake School District 1-3) contains a very brief section on evaluation.  This is the only place that I was able to find any information on technology evaluation for my school.  It isn’t very specific, but contains four questions that will be evaluated, without stating how the evaluation will be conducted.  They are very broad questions about teachers’ skills, student benefits through technology, community benefits through technology, and future needs for improvement.  The evaluation section is very vague as to when it will be done, and it doesn’t contain a clear direction for the evaluation to take place, whether through benchmarks, surveys, interviews, test scores, or a mixture of these.  It doesn’t state how technology can be separated from other content areas that it has been integrated with.  I am not aware of any technology evaluation even taking place, so this is an area that definitely needs work. 

 

Critical Issue 2:  “Current practices for evaluating the impact of technology in education need broadening.” 

     I agree with the authors that finding ways to identify and measure skills and knowledge students gain from using technology is a very difficult task.  It is compounded by the fact that policymakers need summative evaluations showing concrete improvements that students gain from using technology, while educators need formative evaluations to align technology uses with the learning goals and content standards of the curriculum, curriculum integration methods, and learning processes in order to make decisions for continual improvements.  The constant change in technology and the struggle to keep current also complicates this issue.  I don’t believe that my district is even ready for this step, because we really haven’t even started the process of evaluating technology.  I think that keeping Critical Issue 2 in mind will help us when developing our technology evaluation plan, hopefully making our plan broad enough from the start.

 

Critical Issue 3:  “Standardized test scores offer limited formative information with which to drive the development of a school’s technology program.  Most schools are looking for additional means for collecting useful data for this purpose.”

     To improve technology’s effectiveness in schools, they need formative evaluations that tell what works, under what conditions, and with which students.  To me that sounds like a very difficult task, but the authors’ suggestions in the article for using existing data would be considerably helpful, as well as using observations, questionnaires, and standardized tests.  I don’t believe that my district has even started collecting data, but the suggestions offered in the article to look for evidence of technology effectiveness through fewer disciplinary referrals, students’ completing more complex homework assignments, a new robustness in student performances, students taking more difficult electives, etc. will be a great place to start collecting useful data for this purpose.

 

Critical Issue 4:  “Schools must document and report their evaluation findings in ways that satisfy diverse stakeholders’ need to know.”

     Parents, teachers, administrators, funders, policymakers, and taxpayers each have different interests in the effectiveness of technology.  As the article stated, “The best hope of closing this gap lies in helping all stakeholders to see (1) the importance of technology as an effective component of the educational system, (2) how technology is and isn’t capable of making a difference in curriculum and instruction, and (3) how innovative practices of teaching and learning with technology require multiple measures in order to verify its impact.”  Comparing instructional practices and learning opportunities that students have with technology to instruction without technology might be helpful.  Interpreting and reporting data in understandable terms for a variety of audiences could influence the way people think about technology.  In my opinion, this is a major point we will need to consider when reporting data, especially to parents.  It could mean the difference between their support for technology or not.

 

Critical Issue 5:  “In order for evaluation efforts to provide stakeholders with answers to their questions about the effectiveness of technology in education, everyone must agree on a common language and standards of practice for measuring how schools achieve that end.”

     Dialog among stakeholders is very important in evaluating technology.  Therefore, there is a need to come to a consensus on the terms and language of the evaluation process.  Different terms elicit different meanings from different educators, which could harm evaluation efforts.

 

Critical Issue 6:  “The role of the teacher is crucial in evaluating the effectiveness of technology in schools, but the burden of proof is not solely theirs.”

     The teacher is a great source of data about how technology impacts student learning.  What teachers know about their students helps them to make decisions about using technology on a daily basis.  I agree that evaluators need to trust the teacher’s ability to determine and describe technology effects.  I believe that teachers should act as partners with researchers and also help with measuring and documenting changes in student learning because of technology.  I also believe that most of the teachers in my district would be willing to help in the documentation process.

 

Critical Issue 7:  “Because today’s classrooms are expected to be technologically up to date, policies need to be updated continuously to support best practices in technology use.”

     As I read over this section of the article, two issues stuck in my mind.  The first is “What is a school’s responsibility for out-of-school computer access?”  As more and more teachers become computer literate, they are assigning more work requiring the use of computers.  Many of our students do not have access to computers/Internet at home, so the work must be completed in school.  We do have open labs one night a week, and students could make arrangements with teachers to work after school in the labs.  The other issue deals with using email communications with parents.  I agree that frequent communications with parents is positive, but I also worry that some parents may over-use this feature, making it very time consuming for the teacher.

 

Summary:  I found the issues in this article to be very realistic and identifiable.  I hadn’t realized until now that my school district didn’t have a very thorough or effective evaluation plan.  The article had many feasible starting points to help us when creating our technology evaluation plan.  I can see that creating the plan will not be an easy task, but using some of the suggestions laid out in the article should prove to be very helpful.