ooligo
claude-skill

Structured interview loop builder with Claude

Difficulty
intermediate
Setup time
30min
For
recruiter · hiring-manager · talent-acquisition
Recruiting & TA

Stack

A Claude Skill that takes a job description, the role’s level, the must-have competencies, and the eligible interviewer pool with each interviewer’s calibrated strengths, and produces a complete loop design — stage progression, per-stage rubric with anchor descriptions, behavioral questions per dimension, and an interviewer-assignment table with the rationale for each pick. Then stops at a hiring-manager review gate before anything is configured in the ATS. Replaces “we’ll figure out the loop when a candidate is in screen” with a 30-minute design pass that produces operational discipline.

When to use

  • You have an approved JD, a confirmed level, and a list of must-have competencies that differentiate hire from no-hire on this role.
  • You have a structured interviewing rubric library with anchor descriptions per score level per level band. The competency template at apps/web/public/artifacts/interview-loop-builder-skill/references/1-competency-library.md shows the shape; if you cannot fill it in, you do not yet have a rubric this skill can pull from.
  • You have an interviewer pool with calibrated strengths recorded per competency per level band — see references/2-interviewer-strengths.md in the bundle for the matrix.
  • A hiring manager will review the loop before it is configured in Ashby or Greenhouse. The skill writes files and stops; it does not push to the ATS.

When NOT to use

  • Auto-scheduling. This skill designs the loop. It does not schedule interviews, match calendars, or send candidate-facing booking links. That is Goodtime, Ashby Scheduling, or Greenhouse Scheduling. Coupling design and scheduling in one skill couples two failure modes that should fail independently.
  • Replacing the rubric design with the hiring manager. The skill emits anchor descriptions per score level by pulling from the competency library, but the library itself — what a 5 looks like for systems-design at IC5 — is owned by the hiring manager and the head of function. If the library is empty or all-template, the skill refuses and surfaces a TODO rather than inventing rubric anchors for a function it has no calibrated signal on.
  • Generic templated loops without role-specific calibration. If the inputs do not name the level, the must-have competencies, or the eligible interviewer pool, the skill refuses. A four-stage loop with generic “behavioral”, “technical”, “system design”, “leadership” labels reads as structured but is not. Every candidate gets the same questions regardless of role priorities, which defeats the point of structure.
  • Roles below a defined complexity threshold. A two-week contractor role does not need a four-stage loop. The skill warns and suggests a one-stage screen if the role is contract, hourly, or under 6 months expected tenure.
  • Replacing behavioral interviewing training. The questions emitted by the skill follow the situation/behavior/outcome shape, but interviewers still need trained calibration to score consistently. The skill is the scaffold; the training is the prerequisite.

Setup

  1. Drop the bundle. Place apps/web/public/artifacts/interview-loop-builder-skill/SKILL.md into your Claude Code skills directory (or claude.ai custom Skills). The skill exposes one callable function: design_loop.
  2. Fill in the competency library. Copy references/1-competency-library.md to your team repo. Replace every placeholder with your real competencies, definitions, level bands covered, and anchor descriptions per score level. The skill refuses to run if the library is template-only.
  3. Fill in the interviewer-strengths matrix. Copy references/2-interviewer-strengths.md. List each eligible interviewer, their team, and the level bands they are calibrated to score each competency on. The “Last interview” column is the trigger to re-calibrate at 6 months idle.
  4. Configure inputs per role. For a given role, pass the JD path, the level, an array of 3-6 competency IDs, and a path to the filled interviewer-strengths matrix. The skill emits loop.md and per-stage scorecard scaffolds under scorecards/.
  5. Dry-run on a closed loop. Run on a role you designed manually in the last quarter. Compare the skill’s stage mapping and interviewer assignments to the manual design. If they diverge, the competency library or interviewer matrix is usually the thing that needs tuning, not the skill prompt.

What the skill actually does

Six steps, in order. The order matters: deterministic validation and mapping happen before the LLM generates rubric anchors and questions, and the candidate-experience pass at the end re-reads the assembled loop to catch overload that is invisible while assigning each stage in isolation.

  1. Validate inputs. Each competency ID exists in the library; the interviewer pool has at least one calibrated person per must-have competency at the role’s level; the level falls inside the library’s covered bands. Halt with explicit TODOs if any check fails. Designing a Director loop with an IC-only library produces inflated rubrics — this is the step that prevents it.
  2. Map competencies to stages. Recruiter screen evaluates fit and basics (never on-rubric). HM screen takes the top 1-2 differentiating competencies. On-site loop spreads the rest one-per-interview where possible. The one-competency-per-interview rule is opinionated — bundling two competencies into a 60-minute interview produces shallower signal on both, and it makes the rubric harder to apply in the moment.
  3. Generate the per-stage rubric. For each post-screen stage, pull the anchor descriptions for the candidate’s level band from the competency library. Generate 3-5 behavioral questions per dimension following the situation/behavior/outcome shape, plus one suggested probing follow-up per question. Hypothetical “what would you do” questions are excluded by default — they reward articulate guessing over evidenced experience.
  4. Assign interviewers with rationale. For each post-screen stage, propose 1-3 interviewers from the pool. Match by calibration fit (hard requirement), load (no interviewer in more than one stage of the same loop), and diversity of perspective (at least one interviewer outside the hiring team where the pool allows). Each assignment ships with an explicit rationale string.
  5. Candidate-experience pass. Re-read the assembled loop. Total active interview time over 5 hours for IC or 7 for leadership → flag and suggest a take-home. More than 6 distinct interviewers → flag loop fatigue. Two stages probing the same competency → flag redundant signal. Cross-timezone stages without an accommodation → surface a TODO.
  6. Hiring-manager review gate. Write loop.md and scorecards/<NN>-<stage-id>.md. Stop. The skill defines no “publish to ATS” action. The HM opens the file, edits, and configures the loop in Ashby or Greenhouse themselves.

The literal output format and scorecard scaffold layout live in references/3-loop-output-format.md in the bundle. The format is fixed because downstream consumers — interviewer reading the scorecard, debrief facilitator collating scores — need predictable structure.

Cost reality

Per loop design, on Claude Sonnet 4.5:

  • LLM tokens — typically 30-60k input tokens (JD plus competency library plus interviewer matrix plus skill instructions) and 10-20k output tokens (loop plus 3-5 scorecard scaffolds with anchors and questions). On Sonnet 4.5 that is roughly $0.20-0.40 per loop design. A function hiring 8 roles a quarter spends under $5 in model cost on this skill.
  • Recruiter and hiring-manager time — the win lives here. A manual loop design from scratch with calibrated rubric pulls is 90-120 minutes of HM plus recruiter time on the design call, another 30-60 minutes documenting questions and assignments. The skill compresses that to a 30-minute review pass on the generated loop. Per role, that is roughly 90 minutes of senior IC or manager time saved.
  • Setup time — 30 minutes per role once the competency library and interviewer matrix are filled in. The library and matrix are the prerequisite — net-new, those take a calibration session per competency band, which is a structured interviewing investment, not this skill’s setup.
  • Compounding benefit — structured loops produce better quality of hire than ad-hoc loops on every published study going back twenty years. The skill’s win is making “structured” the default rather than the exception by removing the per-role design overhead.

Success metric

Track three numbers per role per quarter, in the ATS:

  • Loop design lead time — hours from “role approved” to “loop configured in ATS”. Should drop materially after the skill is in the loop. If it does not, the bottleneck is HM review, not design — surface the loop earlier in the role-kickoff sequence.
  • Inter-rater agreement on the rubric — per competency dimension, how often do interviewers’ independent scores fall within one point. Should hit 80% or above on calibrated competencies. Below that, the anchor descriptions in the competency library are the thing to tune, not the skill.
  • Quality of hire at 12 months — the long-arc metric the loop is meant to move. Compare cohorts hired through skill-designed loops vs ad-hoc loops on the same role family. If the skill-designed cohort does not outperform, re-examine the competency-to-stage mapping rather than abandoning structure.

vs alternatives

  • vs Ashby’s structured interviewing templates — Ashby owns the configured loop, scorecard rendering, and debrief in one product. Pick Ashby’s templates if you want a managed UX and your team will live in the ATS. Pick this skill if you want the rubric anchors, interviewer-strength matrix, and competency-to-stage mapping in your own repo, version-controlled, with the design step swappable as the competency library evolves. The skill’s output is the input to Ashby’s loop configuration, not a replacement for it.
  • vs generic templated loops — every ATS ships default four-stage templates (“phone screen, HM screen, technical, on-site panel”). They pass for structured at first glance but are not. The same template gets applied to a Backend IC4 and a CS Manager M2, with the same generic questions, regardless of which competencies actually differentiate hire from no-hire on each role. The skill earns its 30 minutes of setup on the second role because the design is per-role-calibrated rather than one-size-fits-all.
  • vs hiring-manager DIY loop design — a senior HM can design a good loop from scratch in 90-120 minutes. They tend not to, because under deadline pressure they reuse the last loop they ran, regardless of fit. The skill’s win is not “designs better than an experienced HM at peak”; it is “designs as well as an experienced HM consistently, across all roles and all weeks”. The consistency is the compounding benefit.
  • vs no structured loop at all — the published meta-analyses on structured interviewing put structured interviews at roughly twice the predictive validity of unstructured ones for job performance. If your status quo is unstructured, the skill is not the question — adopting structure is. The skill is how to make structure cheap enough to actually ship on every role.

Watch-outs

  • Interviewer overload from the same person being assigned everywhere. Guard: the assignment step in the skill enforces “no interviewer in more than one stage of the same loop” as a hard rule. The assignment table surfaces a backup interviewer per stage so the recruiter has a fallback when the primary is unavailable, rather than re-using the primary in two stages.
  • Redundant signal across stages. Guard: the candidate-experience pass re-reads the assembled loop and flags any competency probed in more than one stage. The competency-to-stage mapping table at the top of the loop output makes redundancy visible to the hiring manager at review time.
  • Candidate experience neglected. Guard: the candidate-experience pass is a separate, named step in the skill rather than a sentence at the bottom of the loop. It enforces total-time caps (5 hours IC, 7 hours leadership), distinct- interviewer caps (6), take-home suggestions for competencies that bloat the loop, and timezone accommodation TODOs. Without that pass, “one more 30-minute conversation” accumulates invisibly.
  • Calibration drift inside a single loop. Guard: the rubric block emitted per stage includes anchor descriptions per score level pulled from the competency library, not free-text “rate 1 to 5”. Anchors are the thing that holds calibration when the same candidate is scored by four different interviewers in the same loop. Vague rubric → vague scores → debrief that re-litigates every dimension by anecdote.
  • Hiring manager rubber-stamps the design. Guard: the skill stops at the review gate and writes to files. There is no “publish to ATS” action. The HM has to open the file and edit it before configuring the loop — that friction is intentional. If HMs start signing off without reading, the loop content drifts away from role priorities and the skill stops earning its time saved.
  • Stale interviewer calibration. Guard: the interviewer matrix has a “Last interview” column. Cells aged over 6 months trigger re-calibration before the interviewer is assigned again. When interview intelligence reveals questions that are not producing useful signal, update the competency library’s anchors and the skill emits the new anchors on the next run.

Stack

The skill bundle lives at apps/web/public/artifacts/interview-loop-builder-skill/ and contains:

  • SKILL.md — the skill definition with frontmatter, when-to-invoke rules, inputs, method, and watch-outs paired with guards
  • references/1-competency-library.md — the competency taxonomy with anchor descriptions per score level per level band; fill in per function before running
  • references/2-interviewer-strengths.md — the eligible interviewer pool matrix with calibrated coverage per competency; fill in per team before running
  • references/3-loop-output-format.md — the literal Markdown format the skill emits, including scorecard scaffold layout

Tools the workflow assumes you already use: Claude (the model), Ashby or Greenhouse (the ATS where the HM configures the designed loop), BrightHire or Metaview (interview intelligence whose signal feeds back into the competency library’s anchor tuning). Pairs directly with the JD writer upstream and the interview debrief summary downstream.

Files in this artifact

Download all (.zip)