ID Case File #2 - The Leaky Pipeline

July 28, 2025

The Dilemma

How do you find the root cause of a problem when you can't talk to the people who are actually experiencing it? I'm forwarding you an email from a new lead, the Dean at Northwood University. Take a look…

How do you find the root cause of a problem when you can't talk to the people who are actually experiencing it? I'm forwarding you an email from a new lead, the Dean at Northwood University. Take a look…

A Director of Design explains the expectations for a senior instructional design role, focusing on judgment over technical skill.
A Director of Design explains the expectations for a senior instructional design role, focusing on judgment over technical skill.

An email forward appears in your inbox:

---------- Forwarded message ---------

From: Dr. Evelyn Reed <ereed@northwood.edu>
Date: Mon, Jul 28, 2025 at 10:44 AM
Subject: Urgent Consultation Request
To: Skye Calloway <skye@id.inc>

Dear Skye

For the last four years, our introductory chemistry course, CHEM 101, has become a significant roadblock for our students. It's a required gateway course for nearly all our STEM majors, but we're losing almost half of the students who take it; our DFW rate is at an unacceptable 40%.

The prevailing sentiment among our chemistry faculty is that the problem is simply one of student preparedness. Their consistent recommendation has been to add more tutoring and supplemental instruction. We've invested heavily in these resources, but the needle hasn't moved.

I know the timing is not ideal. It's finals week, which means direct access to students for interviews is impossible, and the faculty are swamped. However, we can provide full access to all of our historical course data, past student evaluations, as well as the course itself.

The faculty will have dedicated time over the upcoming summer break to work with your team to make any necessary changes to the course. To make the most of their time, we need your team to find the root cause now so we can hit the ground running and have the course updated for the fall.

n instructional designer considering his design philosophy, combining Human-Centered Design (HCD) with Andragogy (Adult Learning Theory).
n instructional designer considering his design philosophy, combining Human-Centered Design (HCD) with Andragogy (Adult Learning Theory).


Dr. Evelyn Reed

Dean, College of Sciences
Northwood University

As you can see, it's a classic 'leaky pipeline' problem, but the real challenge is that it's the last week of the semester. We can’t interview students or faculty and, even more importantly, the students who have already failed or dropped the course (the people we really need to talk to) are no longer enrolled and effectively unreachable.

The Dean has given us full access to their systems, but we need to find the root cause without talking to anyone directly.

I’ve scheduled a follow up meeting next week to review our initial findings, so you’ll need to be strategic about where to focus your efforts.

n instructional designer considering his design philosophy, combining Human-Centered Design (HCD) with Andragogy (Adult Learning Theory).
n instructional designer considering his design philosophy, combining Human-Centered Design (HCD) with Andragogy (Adult Learning Theory).

The Decision

As I see it, you have two primary paths you can take for this initial analysis:

Course Design & Analytics:

Dedicate your week to a deep, forensic analysis of the existing course materials and historical student performance data. Dig into their LMS and review everything (syllabi, modules, assignments, and exams) to find patterns in the course design that might be causing students to fail.

Student Feedback & UX:

Prioritize gathering insights from existing student feedback. Review past course evaluations and any university-wide surveys on student experience. Conduct a thorough audit of the online learning environment itself (its usability, accessibility, and clarity) to uncover systemic barriers.

Student Feedback & UX:

Prioritize gathering insights from existing student feedback. Review past course evaluations and any university-wide surveys on student experience. Conduct a thorough audit of the online learning environment itself (its usability, accessibility, and clarity) to uncover systemic barriers.

Select an option above or scroll down to view the debrief.

The Debrief

Both analytical paths led to a positive reaction from the Dean - there is no 'wrong' answer here. The path you chose didn't determine if you found a problem; it determined what kind of problem you found.

Focusing on the course alignment uncovered a clear, data-backed instructional problem: an assessment misalignment. This is a tangible, solvable issue that the faculty can address. It's a very successful and valuable finding.

Analyzing the context and environment of the course uncovered a powerful, human-centered experiential problem: a confusing and unsupportive learning environment. This is a more systemic issue that speaks to the students' lived reality.

The real skill isn't just finding a problem. It's about knowing how to prioritize your analysis to find the root cause. To understand that, we need to look at the full framework we use for any comprehensive Needs Assessment.

Our design process is always grounded in a comprehensive Needs Assessment, which is the systematic process of identifying the gap between the current state and the desired state. In a project with no constraints, we would analyze all four layers. But with such a short turnaround time for our analysis, we have to prioritize. To understand that choice, we first need to look at the four layers of analysis we use.

Task Needs Assessment

A Task Needs Assessment focuses on understanding the specific tasks and skills required to perform a job or, in this case, succeed in a course. We deconstruct the work to find out what knowledge, skills, attitudes, and behaviors (KSAB) are required for effective performance.

This could involve:

  • Analyzing job descriptions and competency frameworks.

  • Breaking down complex tasks into smaller, more manageable steps.

  • Observing experts to deconstruct their intuitive skills.

Reviewing the course design and alignment is a classic Task Analysis. You would be reviewing the syllabus, assignments, and exams to map out every task a student must perform to pass. A thorough analysis here could reveal that an exam, for example, is testing a skill that was never actually taught, creating a clear instructional gap.

Organizational Needs Assessment

An Organizational Needs Assessment aims to align any potential solution with the broader business objectives and strategic goals of the client. It seeks to answer the question: How can our work support the organization's success?

This might involve analyzing: 

  • Strategic goals and initiatives, like new product launches or market expansions.

  • Performance gaps, like low productivity or high safety incidents.

  • External factors, like changes in industry regulations or new market competition.

In this case, the Dean has given us a very clear top-level strategic goal: improve student progression and retention by reducing the 40% DFW rate in CHEM 101. However, a full organizational analysis also involves investigating how the current solution aligns with that goal. A key part of our analysis would be to determine if the course's stated objectives and curriculum are truly designed to support student success or if they are misaligned, perhaps focusing on "weeding out" students rather than building them up.

Learner Needs Assessment

A Learner Needs Assessment is all about understanding the learners themselves: their demographics, backgrounds, motivations, challenges, and learning preferences. Without this layer, we risk creating a solution that is technically correct but completely disconnected from the people who need to take it.

This assessment would analyze: 

  • Demographics and cultural backgrounds.

  • Prior knowledge and existing skill levels.

  • Intrinsic and extrinsic motivations for learning.

Since we can't interview students directly, we would analyze the data they've left behind, like past course evaluations, to build a picture of their experience. We'd look for recurring themes in their feedback to uncover their specific pain points.

Environmental Needs Assessment

An Environmental Needs Assessment evaluates the technological, logistical, and cultural factors that can support or hinder learning.

This might involve:

  • Analyzing the available technological infrastructure, like the LMS or internet connectivity.

  • Assessing the physical learning environment for on-site training.

  • Considering cultural and logistical factors, like organizational culture or time constraints.

For a hybrid course like CHEM 101, an environmental audit might reveal that the LMS is difficult to navigate or that critical resources are buried. These environmental barriers can cause students to fail, regardless of how well-prepared they are.

n instructional designer considering his design philosophy, combining Human-Centered Design (HCD) with Andragogy (Adult Learning Theory).
n instructional designer considering his design philosophy, combining Human-Centered Design (HCD) with Andragogy (Adult Learning Theory).

Deconstructing the Approaches

Now, let's look at the two approaches through that four-layer lens. Both are valid strategies a designer might take, and both have significant pros and cons in this specific situation.

Looking Inside-Out

Analyzing the course and historical data is an 'inside-out' approach. It starts from the perspective of the institution. A core part of this approach is conducting a Task Needs Assessment to ensure alignment. You would analyze if the final exams are truly aligned with the course's learning objectives, and if the instructional materials are aligned with what's being tested. A thorough analysis here could reveal a critical flaw—for example, that the exams cover content that was never actually taught in the online lectures. This path is excellent for finding these kinds of objective instructional gaps.

So, why isn't this the clear first choice? Because of the context the Dean gave us. The fact that the university has already invested heavily in tutoring and supplemental instruction, and it hasn't worked, is a massive clue. It suggests that the issue might not be a simple instructional gap that more 'help' can fix. While this path could uncover the problem, you risk spending your entire week analyzing the curriculum only to confirm what the failed tutoring already implies: that the problem lies elsewhere.

Looking Outside-In

On the other hand, analyzing student feedback and the user experience is an 'outside-in' approach, rooted in our Human-Centered Design philosophy. It starts from the perspective of the learner. By reviewing past course evaluations, you are conducting a Learner Needs Assessment. By auditing the online learning platform, you are conducting an Environmental Needs Assessment.

However, let's be realistic, this approach has its own serious flaws. We can't let our belief in empathy blind us to the data's limitations. Student evaluations are not a perfect source of truth. They are often skewed toward the extremes, the students who loved the course or hated it, and they completely miss the voices of the students who withdrew before the end of the semester. So, we know going in that this data is incomplete.

Making the Best Choice

So, why prioritize this approach? Because in a situation with limited time and a 'black box' problem, our goal isn't to find the definitive answer in one week. Our goal is to form the strongest possible hypothesis. The open-ended comments in course evaluations are a goldmine of qualitative data. They can provide clues about hidden frustrations, like a confusing LMS or a lack of instructor presence. Systemic issues like poor usability or inaccessible materials can create significant barriers. If students struggle to navigate the online environment, they may fail regardless of the content quality, making the environment itself a potential root cause worth investigating.

The Bottom Line

This "outside-in" approach, while imperfect, is a strategic bet that the student's lived experience will give us the clues we need to conduct a much more efficient and targeted Task Analysis later. 

Ultimately, both paths require you to analyze data, but the real job of an instructional designer isn't just to analyze data; it's to find the story hidden within it. That story is what allows you to move beyond the surface-level symptoms and solve the right problem.

n instructional designer considering his design philosophy, combining Human-Centered Design (HCD) with Andragogy (Adult Learning Theory).
n instructional designer considering his design philosophy, combining Human-Centered Design (HCD) with Andragogy (Adult Learning Theory).

Contact us

Get in touch with our expert team

An overhead view of a successful client meeting, where our expert instructional design team finalizes a strategic partnership with a handshake over the conference table.

Contact us

Get in touch with our expert team

An overhead view of a successful client meeting, where our expert instructional design team finalizes a strategic partnership with a handshake over the conference table.

Contact us

Get in touch with our expert team

An overhead view of a successful client meeting, where our expert instructional design team finalizes a strategic partnership with a handshake over the conference table.