M.M.

Mandy McLean, PhD

This portfolio is shared by request. Enter your password, or email to request access.

Need access? Email mndmclean@gmail.com.

Mixed-methods UX research leader.

I build the systems that help organizations turn user understanding into better product and learning decisions at scale.

PhD, STEM Education · Mixed Methods · UXR
Mandy McLean
About

How I work.

I'm a mixed-methods UX research leader focused on building the systems that help organizations turn user understanding into better product and learning decisions at scale.

Most recently, I built and led the UX research function at Guild, scaling it from 1 to 9 researchers and supporting 100,000+ working adult learners across product, design, and strategy. My focus was on building research infrastructure, standards, and enablement that allowed research to operate across teams.

The throughline in my work is systems: how research is structured, shared, and used inside organizations.

At Guild, I also built AI transformation across research and product workflows, including AI-enabled research operations, governance for responsible use, and company-wide enablement through training and practice.

Previously, I completed my PhD in STEM education where I studied how people learn and how learning systems shape participation and outcomes.

Today, I work with education organizations and school districts on AI and learning systems, focusing on how AI is changing instruction, assessment, and learner behavior.

Selected work

A leadership narrative and three case studies.

Click any section to expand. Confidentiality has been respected throughout — Guild proprietary details have been anonymized or generalized.

Leadership

Democratizing research at Guild.

As Guild's first researcher, I built the templates, training, and tiered governance that let teams across the company run their own tactical research while the central team focused on the studies that actually needed it. The team grew from 1 to 9 because the work scaled by reach, not headcount.

Democratization Research ops Team leadership 2018–2025
+
A note on confidentiality. Artifacts and details below have been anonymized or generalized to respect Guild's proprietary information. Frameworks and approaches are shared in their general form; specifics can be discussed in conversation.
11
Partner teams running their own research
20+
Studies running per quarter at peak
1 → 9
Researchers on the central team
7 yr
From first researcher to head of function
The question
How does research scale when you can't hire fast enough to keep up with demand?

Context

The problem we were solving.

I joined Guild in 2018 onto the product design team. Guild was early-stage, growing fast, and figuring out what working adult learners actually needed from an education benefit. Research was scattered across functions, with no shared standard for how questions got asked or how answers traveled back into decisions.

I was hired to do research within product design. But as I built credibility on early projects, requests started coming from outside product — marketing, sales, strategy, leaders trying to understand the workforce we were serving. Within a year, the demand had made the case for a dedicated function. By the time I formally moved out of product design and into a research role, I'd been operating as a shared resource across the company for months. The function existed before it had a name.

The choice in front of me was the choice every first researcher faces — become a service desk, or build something that could outgrow my own bandwidth.

Approach

Scale by reach, not just headcount.

Both happened in parallel. The team grew from one to nine over seven years. But the team grew because we leaned into democratization — building templates, training, and tiered governance that let partner teams run their own tactical research. Hiring more researchers wouldn't have kept up with the demand. Empowering more people to do research did.

I made the case for a different model in different rooms over a long stretch of time. There was no executive sponsor handing me runway. The function got built because I kept showing up — to product reviews, strategy offsites, roadmap discussions — with research that mattered to the decision in front of us. The lesson I kept relearning: being in the room mattered more than being right on paper. A perfectly designed study delivered to people who weren't waiting for it changed nothing. A directionally correct insight raised in the moment a decision was being made changed the company.

The system I built

Three layers, working together.
LAYER 01 Templates For people who weren't researchers Research plans Discussion guides Survey design Doc-style reports Retro frameworks LAYER 02 Standards For the team's shared practice Project lifecycle SOPs Qual best practices Survey best practices QA processes Accessibility standards LAYER 03 Curriculum So partner teams could grow with us UX Research 101 Survey Panels 101 Survey Scales 101 "How We Work" series Office hours & coaching RESULT · A research function 11 partner teams could trust and use

Each layer answered a different question. Templates let designers and PMs do their own tactical research without reinventing the basics. Standards kept quality from drifting as the team grew. Curriculum turned partner teams into capable collaborators. The system worked because all three did.

How I prioritized work

Tiered governance: lean vs. rigorous, partner-led vs. central-team-owned.

The methodological tension between "what would the rigorous answer take" and "what does the business need by Friday" was constant. I made peace with it early: rigor isn't a fixed bar; it's calibrated to the consequences of being wrong. Some questions warranted six-month studies. Others needed a tight survey by end of week. Knowing the difference was the work.

I built that judgment into a tiered model that taught partner teams when to call the central team in.

Three tiers, three modes of support
TIER 01 Self-serve Templates, no consult FITS: Usability tests Simple surveys Internal interviews Retros Run by partner teams TIER 02 Consulted We review and coach FITS: Multi-method studies Segmentation Deeper qual Concept testing Co-led with central team TIER 03 Central-team-owned High stakes, high rigor FITS: Longitudinal studies Public-facing reports Strategic foundational Cross-product synthesis Led by central team
The model taught partner teams to ask "which tier is this?" before "who runs it?"

The tiers weren't bureaucracy. They were a shared vocabulary that let a designer or PM walk into a planning conversation and self-assess whether what they needed was something they could run alone, something they should call in for help on, or something only the central team could do well. It moved the methodological judgment closer to the work, and freed the central team to focus where it mattered most.

Partner teams

The function reached across the company.
11 partner teams at exit
Marketing Product Product Design Engineering Academic Partnerships Employer Partnerships Strategy Business Operations Member Services Coaching Partner Enablement

Impact

What changed.
Before
Research scattered across functions. No shared standard for how questions got asked. Every study a one-off. Demand always exceeded the bandwidth of whoever was doing it. "Can you talk to a few users?" was the highest expectation.
After
Tier 1 work — usability tests, simple surveys, retros — became routine partner-team practice. Tier 2 and 3 work remained central, freeing the team to focus on the studies that needed it. Research stopped being the bottleneck for tactical work.

By 2024, partner teams across the company were running their own usability tests, surveys, and discovery work as routine practice. The central team's nine researchers could focus on the longitudinal, cross-product, and high-stakes studies — including the ones in Case Studies 02 and 03 — that genuinely required their depth.

Several of the people I hired and grew on this team have gone on to lead research at other companies. That's the impact metric I'm proudest of.

01
Case Study

Leading AI transformation in research and across Guild.

Starting in early 2023, weeks after ChatGPT's launch, I led the company's adoption of generative AI as both a research operations capability and a workforce-wide transformation.

AI transformation Research ops Governance 2023–2025
+
A note on confidentiality. Specific tools, vendors, and internal program details have been generalized to respect Guild's proprietary information.
46+
Weekly newsletter editions
18 mo
Sustained company-wide cadence
100%
Research team using AI by 2024
11
Partner teams reached
The question
How do you adopt AI early — without adopting it badly?

Context

Early 2023.

ChatGPT launched in late November 2022. By January 2023 I was already integrating it into my own research workflows — synthesis, lit review, prompt design for survey item generation, faster pattern-finding across qualitative data — and quietly testing where it actually helped versus where it produced confident nonsense. Most of the company hadn't started yet. There was no AI policy, no enterprise tooling, no shared sense of what was acceptable use.

I wasn't asked to lead AI work. I started doing it because it was already changing my own job, and because the questions colleagues were bringing me — "Can I use this for X? Is this risky? What tool should I pick?" — needed answers nobody else was set up to give. By 2024, "AI Transformation" was part of my title. The role grew around the work, not the other way around.

Approach

Three layers, the same shape as the research function.

I structured the AI work the same way I'd structured the research function — enablement for non-experts, governance for shared practice, and a regular rhythm to keep the company connected to a fast-moving field.

LAYER 01 Enablement For the whole company Weekly AI newsletter Workshops & trainings Office hours Working group Use-case library LAYER 02 Governance Human-in-the-loop, by design Acceptable-use guidance Vendor evaluations Privacy & data handling Output validation Risk review checkpoints LAYER 03 Research ops AI inside the research workflow Synthesis & coding Recruiting & screening Transcription & tagging Repository & search Sentiment & NLP RESULT · A company that adopted AI early, broadly, and responsibly

The newsletter

46+ weekly editions, "The AI Update @ Guild."

The newsletter was the spine of the program. Each week I translated frontier AI research, product releases, and policy debates for a non-technical company audience — Stanford labor papers, OpenAI benchmarks, compute-economics primers, the evolving conversation from "AI safety" to "responsible AI." Every issue followed the same shape: a current development, a plain-language explanation of why it mattered for our work, links to the original sources, and a "what's changing" framing for decisions ahead.

The rhythm
Jan '23 Q3 '23 Q1 '24 Q3 '24 Q1 '25 46+ weekly editions, 18 months sustained Each dot is one edition. Each edition was new.

Reach grew quickly. By steady-state the newsletter was reaching most of the company. People forwarded it to their teams. Other companies asked for copies. It became the artifact people associated with Guild's AI culture.

Artifact

Newsletter screenshot. Drop in a representative issue here.

Human-in-the-loop, in practice

Where governance had to live in the daily work.

Enablement without governance is irresponsible. The fastest way to lose trust in AI inside a company is to let an early, confident, wrong AI output reach a customer, a partner, or a board. The role I held wasn't just to evangelize AI; it was to be the person who could tell when an AI output was credible and when it wasn't, and to teach others to do the same.

Mandy — fill in: 1-2 paragraphs about a specific moment where you caught an AI-generated mistake before it shipped, OR a specific governance decision that prevented a problem.
The value of AI in a research function isn't speed. It's speed plus rigor.

Inside the research function

What changed in how we worked.
Before AI
Transcription took days. Coding interviews was manual and slow. Synthesis across long-running studies relied on memory and notes. Survey item generation, screener drafting, and recruiting outreach were each their own time block.
With AI, validated
Transcription and initial coding moved from days to hours. Researchers could query against months of accumulated qualitative data. Item generation, screeners, and outreach got faster. Every speed-up paired with a validation step — built in by default.

I worked with the team to develop a working pattern for AI-assisted synthesis. AI surfaces candidate themes; the researcher tests them against the raw data; the researcher writes the final claim.

The AI never gets the last word. The human never starts from scratch.

Outcomes

What this looked like 18 months in.

The work that mattered most was the rhythm — a weekly artifact that signaled the company was paying attention, a working group that turned questions into shared practice, and a culture that took AI seriously without taking it as gospel.

By the time I left in mid-2025, AI was not a special initiative at Guild. It was part of how the company worked.
02
Case Study

Understanding learner disengagement: a year-long diary study.

A longitudinal mixed-methods study with 12 working adult learners over a full year, designed to surface the moments where motivation and persistence break down.

Longitudinal Diary study Mixed methods 2023–2024
+
A note on confidentiality. Participant quotes, identifying details, and proprietary findings have been anonymized or generalized.
12
Working adult learners
12 mo
Per participant, longitudinal
3
Insights that reshaped strategy
4 teams
Adopted the framework
The question
Why do most paused learners say they'll come back, but only one in four do?

Context

Why we needed a longer lens.

Learner disengagement was a known problem. Most paused learners told us they intended to return; most never did. Fewer than one in four came back, even though over 80% had said they would. The behavioral data could tell us when a learner paused, what they reported at the moment of pause, and which segments paused at higher rates. What it couldn't tell us was the texture of the weeks before a pause.

The moment a learner paused wasn't the moment the pause began.

To redesign anything that mattered — coaching, communications, program structure, support — we needed to understand the trajectory, not just the endpoint.

Approach

A year of weekly diaries.

I designed a year-long video diary study with 12 working adult learners, purposively sampled across employer partners, program types, and demographic segments. The full study ran a calendar year per participant — long enough to capture the seasonal rhythms, life shocks, and recovery cycles that shorter studies kept missing.

Study design
Month 1 Month 4 Month 7 Month 10 Month 12 Weekly video diaries Quarterly 60-min interviews Q1 Q2 Q3 Q4
12 participants × 12 months. Weekly asynchronous video diaries (5–10 min) + quarterly 60-minute interviews.

The methodological choices were deliberate. Single in-depth interviews would have been retrospective. Surveys at scale already existed and told us what, not why. Focus groups would have let social pressure suppress the unguarded disclosure. The diary study traded sample size for time depth — and time depth was what was missing.

Finding

Three insights.
01
Pause is a trajectory, not a discrete decision point.

Disengagement was a cascade that began weeks before a learner disappeared. Across the 12 participants, the shapes of their years were different in detail but recognizably the same in structure: an initial lift, an inevitable dip, and then a fork.

The pattern we found
ASSUMED · V-SHAPE High Low Before Holiday After Quick recovery after the shock OBSERVED · U-SHAPE High Low Before Holiday After Long bottom — being-behind doesn't end with the holiday
Holidays didn't produce the V-shape everyone assumed. They produced a U with a long bottom.
02
The fork is the point of leverage.

Life shocks couldn't be prevented — illness, caregiving, financial stress — but the conditions at the fork could be shaped. By the time a learner paused, they were typically already behind, and being-behind was itself a demotivation signal. We tested the spiral pattern against behavioral data and found the holidays didn't produce the V-shape everyone assumed; they produced a U with a long bottom.

03
Semester one is re-acclimation, not capacity.

Working adults stacked courses in their first term. Coaches encouraged this — "how much can you handle?" was the default. But semester one is where academic re-acclimation happens, especially in gateway math. The capacity frame was wrong: semester one isn't a test of throughput, it's a re-entry into the role of student. Learners who took fewer courses up front persisted at meaningfully higher rates.

Impact

What changed.
Before
Disengagement framed as event. Coaching outreach generic and immediate. Comms used performance language ("don't fall behind"). Course-load decisions framed as throughput. Holiday campaigns timed to the shock itself.
After
Disengagement framed as trajectory. Outreach fork-aligned and learner-specific. Comms used momentum language. Course-load conversations reframed around re-acclimation. Holiday campaigns built around the post-holiday recovery window.

Each insight reshaped a different part of how the company supported learners. Coaches got vocabulary for disengagement patterns they could feel but couldn't name. Comms timing moved from immediate post-pause contact to the weeks after, when learners were actually trying to come back. The pre-existing pause survey was rewritten to feed a real-time retention dashboard. Industry-specific shock calendars (retail's back-to-school, healthcare's shift demands) became standing input to campaign timing.

The spiral framework was adopted across product, coaching, marketing, and strategy as a shared vocabulary for retention.

Disengagement is a gradual erosion of momentum, not a sudden decision. The work that matters most isn't preventing the shock — it's meeting people at the fork.
03
Case Study

When yes isn't enough: rebuilding a stalled funnel.

A mixed-methods diagnostic of a critical funnel step that had been declining year-over-year since 2020. The finding reshaped how Guild engaged learners in a part of the journey nobody had been watching.

Mixed methods Funnel diagnostic UX research 2023
+
A note on confidentiality. Specific funnel metrics, member quotes, and product details have been generalized.
3+ yr
Of declining funnel rate, reversed
8–10
In-depth interviews
100s
Survey responses per month
4 teams
Coordinated to ship the response
The question
Why are people who already said yes not showing up?

Context

A quiet decline.

Guild's funnel had a step between application approval and program start that had been declining year-over-year since 2020. It wasn't a dramatic drop. It was a slow erosion easy to miss in any single quarter. The trend was visible only when I pulled together a longer view of the data during ongoing review.

The puzzle was simple to state and hard to answer. These were people who had cleared eligibility, completed an application, gotten approved. And then a meaningful share never actually started. Existing assumptions ranged from "life got in the way" to "the product was confusing." Neither was specific enough to fix.

The pattern
Eligible Application Approved "Yes" THE GAP Approval → Start Declining since 2020 No engagement built in Started Active learner A meaningful share who said yes never made it to "started."

Approach

Watching the gap from both directions.

I designed a two-track study. A new ongoing touchpoint survey reached members at the moment they hit that funnel step — hundreds per month — capturing self-reported readiness, intent, and barriers in close-to-real-time. Eight to ten in-depth interviews with members who had paused at exactly that point gave us the texture the survey couldn't.

Two-track research design
Track 01 · Quant
Touchpoint survey
Sent at the moment members hit the funnel step. Designed for under-a-minute completion. One open-text field carried most of the signal. Hundreds of responses per month.
Track 02 · Qual
In-depth interviews
8–10 interviews with members who had paused at exactly that funnel step. Qualitative recruiting was informed by what the survey was already showing — the tracks worked together.

The hardest design choice was the touchpoint survey itself. Surveys at moments of decay are notoriously underpowered — the people most likely to disengage are also the least likely to respond. I wrote it to be answerable in under a minute, with one open-text field that turned out to be where most of the signal lived. I had the survey running before the qualitative interviews were scheduled, so the qualitative recruiting could be informed by what the survey was already showing.

Finding

"Yes" decays.

The headline insight was simple and not what anyone expected. Motivation eroded in the gap. Approval wasn't a permanent state. "Yes" was a perishable signal, and without something to keep the commitment alive between approval and start, it quietly decayed.

The mechanism
High Low Motivation Approval Program start Time "Yes" No engagement: motivation decays With engagement: commitment sustained
The gap had no engagement. There was nothing to push against the natural decay of intent.

It wasn't primarily life events. It wasn't primarily product friction. The interviews showed people who had been excited at approval describing themselves weeks later in language that revealed slipping commitment. Inertia, not friction, was the failure mode.

Members who didn't start were members for whom nothing in particular had happened — and that was the problem.

Resolution

Filling the gap.

The recommendations I made were structural, not tactical. We didn't try to remove the gap — for many members it was unavoidable, driven by enrollment cycles and program calendars. We built engagement into it.

Before
Silent waiting period. Members got an approval email and then nothing until program start, sometimes weeks later. Coaching kicked in only after the start date. Comms were calendar-anchored, not journey-anchored.
After
Structured engagement sequence in the gap. Coaching outreach moved earlier. Comms timing rebuilt around the rhythm of the gap. Next-steps experience inside the product gave members something to do during the wait.

Several teams owned the implementation. Research had diagnosed the problem and named the mechanism, but coaching, comms, and product each held different pieces of the response.

Impact

A measurable lift.

The funnel step rate improved measurably after the changes shipped. The exact numbers are confidential, but the direction was clear and the change held over time. More importantly, the framing — that motivation is perishable in the gaps between commitment and action — became a shared mental model across the teams that owned the journey. Subsequent work on other funnel steps adopted the same lens.

The dangerous moments in a user journey aren't always moments of friction. Sometimes they're moments of nothing.