The Danger of Ignoring Context in Program and Evaluation Design
By Nora F. Murphy Johnson, PhD
In program and evaluation design, the significance of context is often underestimated, if not outright ignored. This oversight, whether intentional or not, can have profound implications. In fields where “evidence-based best practices” are revered, the nuanced realities of varied environments and populations are frequently overlooked in favor of a standardized approach. However, this one-size-fits-all mentality can lead to ineffective, and sometimes harmful, outcomes.
In this blog post, I talk about the critical role context plays in shaping successful interventions and evaluations, and I offer a strategy to make this task more manageable, and I offer a more nuanced, adaptable framework that respects the unique characteristics of each setting and population.
The Consequence of Overlooking Context in Research, Evaluation, and Programming
In the realm of complexity theories, a core tenet is that causality is an elusive concept, not discernible in advance and, sometimes, only somewhat knowable in hindsight. This theory challenges the conventional understanding of cause-and-effect relationships, emphasizing that causal connections and patterns are invariably…