Why this exists
The literature review (see pedagogy.md and research/summaries/) made one thing clear: the empirical evidence base on adult digital-fluency transfer is thin. No RCTs, no mental-model-construction studies, mixed LLM-tutor metacognitive results. What does exist — Urban Institute 2019, RAND 2024 pilot, ProLiteracy syntheses — leans heavily on provider interviews and program-level observations, not controlled trials.
That is a feature, not a bug. The people who actually know what works with low-fluency adult learners are the librarians, ABE instructors, workforce counselors, and senior-program coordinators who have been teaching digital skills for 20+ years. Their knowledge is largely undocumented in the academic literature. Touring those programs is the highest-leverage research move available to us before we lock the spec.
This document defines what we're trying to learn, who we should learn it from, what to ask, what to bring back, and how to convert the findings into product decisions.
The five questions we need to answer
Each question maps to a specific design choice that's currently underspecified in our docs. Field research must produce evidence that lets us make these choices, not just stories.
Q1. Where do adult learners actually fail?
The Urban Institute names "multi-step workflows" and "cross-application navigation" as failure modes, but at the level of generality that doesn't help us design a curriculum. At the level of specific micro-tasks, where do real adult learners get stuck?
Examples of what we want to surface:
- Is the failure usually at the concept level (didn't understand what a folder is) or the interaction level (didn't know to right-click)?
- Which apps do people most frequently need help with — and which are they too embarrassed to ask about?
- What's the most common reason a learner abandons a task partway through?
- What does "stuck" look like behaviorally — and how do experienced instructors detect it before the learner gives up?
This is the content design question. We can't pick the right tasks for our curriculum without it.
Q2. What does the AI co-pilot actually need to do — and what should it never do?
Our pedagogy doc commits to specific moves (refuse-until-effort, explicit pattern naming, metacognitive debrief, contrasting cases). Provider experience is the reality check on these.
- When does an experienced instructor intervene vs. let a learner struggle? What signal triggers them?
- What kinds of help are most welcomed — and which ones are condescending or demotivating?
- How do instructors handle the user who just wants the answer? Do they ever give it? What changes their behavior between learners?
- What language do they use — and what do they avoid? ("Don't use the word 'simply.'")
- How much praise is too much? How much is not enough?
This is the AI co-pilot design question. The technical-approach doc commits to a specific intervention pattern that the field could either validate or reshape.
Q3. What does engagement look like over time, and what kills it?
Bastani et al. found engagement (not problem volume) mediates learning gains, but their population was high-school students in a structured course. Adult learners self-select in and out continuously. Sustained engagement is harder.
- How long does the average learner stay in a digital-skills program before dropping out?
- What's the longest sustained engagement providers have seen, and what produced it?
- Is in-person component necessary? The Urban Institute brief says yes — providers emphasize "humanizing" the experience. What specifically does that mean operationally?
- What role does cohort / peer learning play? Is the loneliness of self-paced online learning a structural defeat?
This is the engagement architecture question. It probably reshapes whether our product is purely individual or has a community/cohort layer.
Q4. Who actually shows up — and who doesn't?
The PIAAC stats name 32M Americans with no digital skills, but the population that actually walks into a library digital-skills class is a self-selected subset. Understanding the gap matters for distribution.
- Who shows up to your programs? Demographics, motivations, prior failures?
- Who you wish would show up but doesn't? What would change that?
- What's the entry path — referral from another social service, walk-in, online sign-up, employer-mandated?
- What's the trigger event — what happened in someone's life that made them seek out training this week?
This is the distribution / market question. It directly informs the pitch's "first-cohort plan" gap.
Q5. What do good outcomes actually look like — and how do you know?
Providers have been doing this work for decades. They have an intuitive sense of which graduates "got it" and which didn't. We need to surface that tacit assessment.
- How do they know if a learner has actually learned something vs. memorized the steps?
- What do graduates do six months later? Twelve months later?
- What's the most common pattern among learners who didn't succeed despite completing the program?
- What outcomes do funders ask for — and how does that distort what gets measured?
This is the assessment design question. Our spec commits to a near/far transfer distinction; the field probably has practical, lower-cost proxies we should adopt.
Who to talk to
Tier the outreach. Start with the people closest to the work.
Tier 1 — Public libraries (the most universal venue)
Libraries are the dominant US infrastructure for free adult digital-skills training. Per the Urban Institute brief, "libraries have been engaged in computer trainings since the 1990s." This is the single most important provider type to study.
- Local branches — your nearest library system. Walk in. Ask if they teach digital skills. Find the person who runs the program.
- Large urban systems with mature programs — NYPL, Brooklyn Public Library, Chicago Public Library, San Francisco Public Library, LA Public Library. They typically have a Digital Inclusion or Tech Goes Home equivalent.
- DPLA (Digital Public Library of America) and PLA (Public Library Association, ALA division) — they have Digital Literacy initiatives and may connect you to a wider provider network.
Specifically ask for:
- Senior librarians who have been doing this for 10+ years (they've watched everything fail).
- "Tech help" desk staff who do 1:1 (they see failure modes vision-only research can't).
- Anyone running an "intro to computers" or "intro to email" class.
Tier 2 — Adult Basic Education (ABE) and ESOL providers
The federal AEFLA-funded ABE network reaches adult learners who lack high school equivalency. Many integrate digital skills into literacy/numeracy work.
- Local community colleges that host ABE programs — most metro areas have one.
- ProLiteracy — national member organization; their resources page lists affiliated providers.
- LiteracyDC (or local equivalent in your city) — direct-service ABE provider.
- World Education / EdTech Center — they specifically work on technology integration in ABE; published the major recent practitioner literature.
Tier 3 — Workforce development and senior programs
- Goodwill local affiliate — they run digital skills training inside workforce development.
- AARP Senior Planet / OATS — programs specifically for older adults; they have decades of data on what works.
- Local workforce development board (every state has them, federally mandated) — they fund and oversee training providers.
- Senior Community Service Employment Program (SCSEP) — federal program for low-income seniors 55+; the Urban Institute brief specifically interviewed an SCSEP director.
Tier 4 — Academic and research
After you've done field interviews, then talk to the researchers who study this. Their input lands differently after you've seen the work yourself.
- World Education Inc. EdTech Center (Jen Vanek, others) — they bridge research and practice.
- EDC (Education Development Center) — adult learning research.
- MDRC — they evaluate workforce programs rigorously; if anyone has done a real adult-digital-skills RCT it's them.
- Stanford CRADLE / Schwartz lab — Daniel Schwartz himself; the contrasting-cases pedagogy lives here. He may not respond to a cold email but a well-framed one might.
- Wharton — Hamsa & Osbert Bastani — the AI tutor RCT authors. Worth one outreach attempt.
What to do at each visit
Before the visit
- Read their published materials. Most library systems publish their digital-literacy curriculum (it's often Northstar-aligned). Read it. Don't ask them basic questions about it during the visit.
- Send a short email. "I'm researching how to design a better digital-fluency learning platform. Would you have 30–60 minutes to talk about what you've learned teaching this for X years? I'm not selling anything, I just want to understand what works." Identify yourself plainly. Do not pitch the product.
- Bring a one-pager — what you're working on, in non-product-pitch language. They'll ask. Have it ready, but don't lead with it.
During the visit — the interview
Use a semi-structured interview approach. Have the five questions above as the spine, but follow what's interesting.
The most valuable question is "tell me about a learner who failed." Specific stories surface mechanism in a way general questions never do. Ask for two or three of these. Take notes.
The second most valuable question is "show me the materials you actually use." Often what providers say they do and what they actually do diverge. Worksheets, cheat sheets, screenshots they print out — these are gold.
The third most valuable observation is the room itself. What's on the walls? What's on the whiteboard? Is there a 1:1 tech-help desk and what's the queue look like? Are learners using their own laptops or library desktops? What's the OS — Windows? Chrome? iPad? Each of those choices encodes years of experience.
If they offer to let you sit in on a session, say yes. One observation session is worth five interviews. Take field notes during; don't participate.
After the visit
Write up notes within 24 hours. Format: see template below. Do not skip this — memory degrades fast and you'll do 8–15 of these visits.
Note-taking template
Save each visit as research/field-notes/YYYY-MM-DD-[provider-name].md.
# {{Provider name}} — {{date}}
**Where:** {{city, organization, branch}}
**Who:** {{names, roles, years in role}}
**Format:** {{interview / observation / both, duration}}
**Pre-read:** {{materials I read before going}}
## Top 3 takeaways
1. {{single most important thing I learned}}
2. {{...}}
3. {{...}}
## Failure stories
- {{Specific story of a learner who failed, why, what they tried}}
- {{...}}
## What works
- {{Practices the provider says actually work, with specifics}}
- {{...}}
## Surprises / counter-evidence
- {{Things that contradicted my assumptions before the visit}}
- {{...}}
## Quotes worth keeping
> "..."
## Materials they shared
- {{worksheets, screenshots, link to curriculum}}
## Implications for our design
- {{Specific product/spec/pedagogy implications — one bullet each}}
## Followups
- [ ] {{People they referred me to}}
- [ ] {{Questions I forgot to ask}}
Sequencing — what order to do this in
You probably can't visit 15 providers. You probably should visit 6–8. Here's how to sequence.
Phase 1 (weeks 1–2): Local immersion.
- 2 library visits (your nearest urban branch + one suburban/rural if accessible).
- 1 ABE provider visit.
- 1 senior-program visit (AARP Senior Planet has chapters in many cities; a local senior center is the alternative).
Goal: develop intuition. After this phase you should be able to predict what providers will say. That intuition is what makes the next phase efficient.
Phase 2 (weeks 3–5): Targeted depth.
- 2 visits to programs known for innovation (e.g., NYPL TechConnect, Senior Planet OATS, Goodwill GoodGuides). These are the ones you've heard cited in the literature.
- 1 observation-only session (sit through a 90-minute class, no interview).
Goal: pressure-test the patterns from Phase 1 against the most sophisticated providers. If your Phase 1 hypotheses survive contact with the people who have run the best programs, they're load-bearing.
Phase 3 (weeks 6–7): Researchers and synthesis.
- 1–2 conversations with researchers (World Education EdTech Center, MDRC, possibly Bastani at Wharton).
- Synthesis writeup: what changes in the spec, the pedagogy doc, and the pitch as a result of field research.
Goal: convert field findings into concrete product changes. Researchers help you generalize what you saw without overgeneralizing.
What to bring back
Each visit should produce a field-notes/ markdown. The synthesis at the end of Phase 3 should produce:
- A revised list of failure modes — what real learners actually fail at, prioritized. This rewrites the spec's curriculum and assessment.
- A revised co-pilot intervention spec — what the AI should do, what it should never do, drawn from observed instructor behavior. This rewrites
technical-approach.md's co-pilot section. - An engagement-architecture decision — purely individual, cohort-based, or hybrid. This is a big spec change if it goes anywhere other than individual.
- A distribution-channel hypothesis — D2C / library / workforce / government, with the specific reasoning. This fills the scaffolded TODO in the pitch.
- A list of partner-provider candidates — 2–3 organizations that would plausibly host an MVP pilot. This is what makes the pitch's "first-cohort plan" concrete.
Budget and time
Realistic estimate: 6–8 weeks of part-time effort, ~30–60 hours total. Travel cost minimal if you stay regional (most major cities have all three tiers represented). The single largest hidden cost is writing up notes within 24 hours of each visit — block calendar time for that.
What to skip
- Don't run formal user research with learners themselves at this stage. That comes after the spec is reshaped by provider input. Premature learner research at the spec-defining stage usually produces "users want everything to be easier" findings that don't constrain design.
- Don't try to pilot the product during this research. Two distinct activities. Mixing them is how startups end up with a confirmation-biased view of what their MVP needs.
- Don't try to recruit partners on the first visit. Earn the right to ask by demonstrating you've understood their world.
Success criteria for the program
You're done when:
- You can predict what a new provider will say in the first 10 minutes of an interview, with ~70% accuracy. (You've absorbed the field's tacit knowledge.)
- You have 2–3 concrete spec changes that you would not have made without the field research. (The research actually changed your design, rather than confirming it.)
- You have a credible sentence to put in the pitch that begins "We have visited and learned from N digital-skills providers including [names]." (Distribution credibility.)
- You have 2–3 named providers who would discuss hosting an MVP pilot. (First-cohort path.)
If you finish the program and none of those four are true, the field research has not done what it needed to do — extend or redirect.
How this connects back
Field findings update three docs:
pedagogy.mdgets a new section: "What the field knows that the literature doesn't." Adds practitioner-grounded constraints.product-spec.mdgets revised curriculum, revised co-pilot intervention spec, possibly an engagement-architecture section.pitch-and-overview.mdfills its scaffolded TODO sections (distribution channel, first-cohort plan, founder context if your visits become part of your "why us" story).
The plan we already have (pedagogy → technical → trim pitch & spec) assumes field research happens between drafting pedagogy and drafting technical-approach. That's the right insertion point. The pedagogy doc is grounded in the research literature; the technical-approach doc and the spec revisions should be grounded in the field.