Skip to main content

What I Thought I Knew vs. What I’ve Learned About Evaluation

The Good

Competencies Rated as a Relative Strength 

Competency

Start of Course

End of Course

Notes

2.11. I am able to analyze data using credible, feasible, and culturally appropriate procedures.

4

6

  • analyze the data in depth

  • Triangulation between multiple sources

  • Used the Braun and Clarke Framework

  • Make data analysis more logical with specific procedures

2.14. I am able to use evidence and interpretations to draw conclusions and make judgments and recommendations when appropriate.

3

6

  • comfortable drawing conclusions from data but struggled when data didn’t  match

  • Practiced writing recommendations that held conflicting evidence together

  • build limitations into the recommendations

Competency 2.11:

At the start of this course I rated myself a 3. I could pull themes out of interview data, but if someone had asked me to explain exactly why I organized the data the way I did, I would not have had a clean answer. I was making judgment calls I could not fully name.

That changed when I worked through this evaluation using Braun and Clarke's thematic analysis framework. Before this course, "qualitative coding" to me meant reading through data and grouping things that seemed similar. Braun and Clarke made me realize that is only partially true. Their framework has specific phases: familiarizing yourself with the data, generating initial codes, building themes, then going back and checking whether the themes actually hold up against the raw data. Following those steps in sequence meant I kept catching myself making leaps I would have previously let slide. I would construct a theme, go back to the transcripts to check it, and realize two of the quotes I had grouped together were actually saying different things. Without the framework giving me a reason to go back and check, I would have left that error in.

Running JASP at the same time pushed this further. I had quantitative output sitting next to qualitative themes, and I had to figure out what each one was actually contributing and where they overlapped. That forced a level of precision about what each analysis method can and cannot do that I had never needed before when I was only working with one type of data at a time.

The AI-assisted coding piece added a different kind of challenge. When a tool is generating initial codes for you, the question shifts from "what do I see in this data" to "do I actually agree with what this tool is seeing, and can I justify any place where I override it." That turned out to require more active analytic judgment than coding from scratch, not less, because I was responsible for every decision the tool made that I let stand. Working through that is what moved me to a 6. By the end I could explain every step of my analysis process, defend each decision, and describe exactly where my judgment was doing the work.

Competency 2.14:

At a 4, I was already doing more than restating findings. I could look at data and say something meaningful about what it suggested. What I was not doing well was handling situations where the data did not give me a clean answer, where qualitative and quantitative findings were pulling in different directions, and where the evidence was solid enough to be useful but not solid enough to support a strong universal recommendation.

The data analysis in the evaluation work put me directly in that situation. I had quantitative results showing improvement on one measure alongside qualitative themes where participants were expressing real frustration with the same program component the numbers were praising. Previously my response would’ve been to pick the more compelling narrative and lead with it. What I learned to do instead was write a recommendation that held both things at once. That is a harder sentence to write than "the program is working" or "participants are unhappy," and it is also a more honest and useful one.

The limitations piece was equally concrete. In the evaluation, the data had real constraints: the sample was small, the timeframe was short, and some of the measures were self-reported. Learning to work those limitations into the recommendation itself rather than footnoting them at the end changed how I wrote conclusions entirely. Instead of "the program improved participant outcomes," it became "within the scope of this evaluation, and given the sample size”. That framing is not weaker. It is more accurate, and it gives whoever is reading the report the information they need to decide how much weight to put on it. Getting comfortable writing that way, consistently and without feeling like I was undermining my own work, is what the jump from 4 to 6 actually looked like for me.

The Not So Good:

Competencies Rated as a Growth Area 

Competency

Start of Course

End of Course

Notes

1.8. I am able to identify how evaluation practice can promote social justice and the public good.

1

1

  • understood the principle in theory

  • Did not encounter a clear example this semester of what this looks like when an evaluator does it well in practice,

  • Still uncertain where an evaluator's commitment to justice ends and bias in findings begins

  • feels like a value I am supposed to hold rather than a skill I know how to practice.

1.3.  I am able to select evaluation approaches and theories appropriately.

2

3

  • knew evaluation approaches existed but could not have named them or explained why one would fit better than another.

  • selected an approach because it seemed reasonable aka no reasoning

  • still defaulting to what feels intuitive rather than comparing

  • I need more exposure to different approaches 

Competency 1.8:

I came into this course rating myself a 1 on this competency and I am leaving it the same way, which is itself worth reflecting on. The rating did not move because I am still not sure what this competency is actually asking me to do in practice. I understand the words — evaluation should serve the public good, evaluators should be attentive to equity and justice — but when I try to connect that to concrete decisions I would make inside an evaluation, it gets blurry fast. Which stakeholders count as the public? Whose definition of justice applies when stakeholders disagree? At what point does an evaluator's commitment to social justice start shaping findings in ways that compromise objectivity — and is objectivity even the right goal to begin with? I do not have settled answers to any of those questions, and I think that uncertainty is exactly why the rating stayed at 1.

Competency 1.3:

I moved from a 2 to a 3 on this competency, which is real progress, but the honest version of that progress is that I went from not knowing evaluation approaches existed as a deliberate choice to knowing they exist without yet feeling confident about how to choose between them. In the evaluation work this semester, I selected an approach because it seemed like a reasonable fit for the question we were trying to answer. I did not work through a systematic comparison of alternatives and land on a decision I could fully defend. I picked what seemed to make sense and moved forward.

Right now when I look at an evaluation question I have one or two methods that come to mind. An evaluator at the higher end of this scale would be running through a broader menu and making an explicit case for the choice they land on. I am not there yet, and getting there is going to require more exposure to actual evaluation practice than a single course can provide.

Acknowledgement: Portions of this reflection were developed with the assistance of AI writing tools. All ratings, course experiences, and personal interpretations are my own. AI was used to help articulate and organize my thinking, not to generate the substance of my self-assessment.

Comments

Popular posts from this blog

Blueprints for Better Learning: How Theory Shapes My Design

  Introduction When designing learning experiences, theory isn’t just an academic checklist — it’s the blueprint for creating materials that work. Without a strong theoretical foundation, even the most visually appealing or technologically advanced asset risks falling flat. Learning theories help us answer critical questions: How will learners stay engaged? How will they process the content? What motivates them to apply what they’ve learned?  Over the past five modules, I’ve applied various theories and frameworks from behaviorism and gamification to Self-Determination Theory and the Community of Inquiry (CoI) model to create learning experiences that are engaging, structured, and grounded in research. This post walks through each of those applications, highlighting how theory guided my design choices.

Online and Blended Learning: More Than a Modality

When I think about online and blended learning, I don’t define it by platforms, tools, or whether students are physically present. Instead, I define it by how learning is designed, supported, and experienced. Online and blended learning, at their best, are intentional learning environments that extend access, personalize learning, and challenge traditional assumptions about where and when learning happens. What Does Online and Blended Learning Look Like to Me? To me, online and blended learning look less like a digital replica of a face-to-face classroom and more like a carefully choreographed experience. It includes clear structure, predictable routines, meaningful interaction, and opportunities for reflection. As Conceição and Howles (2020) emphasize, effective online learning environments are intentionally designed with attention to learner engagement, cognitive presence, and instructor presence. In practice, this means learning activities that encourage dialogue (discussion boards ...

Turns Out I've Been Evaluating Things This Whole Time!

A self-assessment that was more revealing than expected If someone asked me to rate myself as an evaluator right now, on a formal scale from 1 to 6, I would land at a 3. Not falsely modest, not overconfident. Just honest about where I actually am versus where I would like to be. I feel grounded in some foundational skills, especially the ones tied to reflection and growth, but I am still developing the more structured, theoretical side of evaluation: program logic, design frameworks, and the discipline of naming the assumptions that sit underneath everything we do. The interesting part of doing this self-assessment was realizing I have been doing informal evaluation for years without calling it that. It turns out that is both encouraging and a little humbling, because "informal" and "rigorous" are not the same thing, and the gap between them is exactly where my growth needs to happen. Where I feel most confident Two competencies stood out immediately because they ar...