Lesson Flow

Learn

Goals and Concepts

Start with the capability target and concept set for this module.

Practice

Studio Activity

Apply the ideas in a guided activity tied to realistic outputs.

Check

Assessment Rubric

Use the rubric to verify competency and identify improvement targets.

Interactive Lab

Practice in short loops: checkpoint quiz, microtask decision, and competency progress tracking.

Checkpoint Quiz

Q1. Which output most clearly demonstrates module competency?

Competency is shown through measurable, method-linked evidence.

Q2. What should always accompany a technical claim in this curriculum?

Every claim should include boundaries and uncertainty.

Q3. What is the best next step after identifying a gap in understanding?

Progress improves when gaps become explicit practice targets.

Glia Boundary Microtask

Which action best reduces glia-neuron boundary errors?

Progress Tracker

State is saved locally in your browser for this module.

0% complete

Glia Identification Annotation

Choose the hotspot with the strongest glial morphology evidence.

Glia identification figure

Selected hotspot: none

Capability target

Execute a proofreading triage cycle that ranks corrections by impact and issues a transparent QC decision.

Hidden curriculum scaffold

Concept set

1) Impact-weighted triage

Not all errors are worth fixing. A merge error on a large interneuron with 500 synapses is far more impactful than a split error on a tiny axon fragment with 2 synapses. Impact factors include: (a) size of the affected segment (larger = more connections affected), (b) position in the analysis region of interest, (c) error type (merges corrupt the graph more directly than splits), (d) confidence of the error detection (is it definitely wrong or ambiguous?).

A practical triage matrix assigns each error a composite score. For example, a merge error on a neuron within the analysis region that has high synapse count and clear evidence of incorrectness scores 4/4 and goes to the top of the queue. A possible boundary error on a small fragment outside the region of interest scores 1/4 and can be deferred or ignored entirely. The key insight is that proofreading time is a limited resource, and spending it on low-impact corrections means higher-impact errors persist in the final dataset. Teams should establish their triage criteria before beginning a proofreading campaign, not invent them ad hoc as errors appear.

2) Proofreading strategies

Each strategy trades off thoroughness against scalability. The choice depends on the scientific question, available labor, and volume size.

In practice, most projects combine strategies. A team might use priority-ranked proofreading across the full volume, then switch to exhaustive local correction within specific subregions that are central to their analysis. The important thing is to document which strategy was applied where, so downstream users know the expected quality level in different parts of the dataset.

3) Quality metrics for release decisions

No single metric captures all aspects of segmentation quality. Use multiple metrics and understand what each one measures.

When metrics disagree – for example, ERL is above threshold but synapse F1 is below – the resolution depends on the scientific question. If the analysis is primarily topological (which neurons connect?), synapse F1 dominates. If the analysis is morphological (what do neurons look like?), ERL matters more. This is why release decisions are scientific judgments, not mechanical threshold checks.

4) Documentation and reproducibility

Every correction should be logged: what was changed, why, when, and by whom. CAVE’s materialization system (see content library) provides this automatically – each edit is a versioned operation, and materialization snapshots freeze the state for reproducible analysis.

Without versioned documentation, two researchers analyzing the “same” dataset at different times may get different results because proofreading edits changed the segmentation between their queries. Materialization versions solve this: you pin your analysis to a specific version number, and anyone can reproduce your results by querying that same version. This is the connectomics equivalent of specifying which genome assembly build you used for a genomics analysis.

Good correction logs also serve a pedagogical function. When new proofreaders join a project, they can review past correction decisions to learn the team’s standards. Entries like “Merged fragments 847291 and 847305 because continuous membrane visible in z-slices 4021-4035; confirmed by checking orthogonal views” teach both the decision criteria and the verification process.

Core workflow

  1. Classify errors by type (merge/split/boundary) and estimated impact (high/medium/low). Use the error taxonomy from the content library as a reference checklist.
  2. Prioritize correction queue: high-impact merges first, then splits in the region of interest, then boundary errors. Defer or discard low-impact errors outside the analysis region.
  3. Apply corrections using Neuroglancer/CAVE split and merge operations. For each correction, note the supervoxel IDs involved and the evidence that motivated the edit.
  4. Verify each correction: check that the fix didn’t introduce new errors. Splitting a merge sometimes creates an orphan fragment that needs re-merging elsewhere. Merging a split sometimes absorbs a nearby fragment that shouldn’t be included. Always inspect the result in at least two orthogonal views.
  5. Record QC decision: compute metrics, compare to release thresholds, issue go/rework recommendation. If the recommendation is “rework,” specify which error categories need further attention and estimate the additional effort required.

60-minute tutorial run-of-show

Pre-class preparation (10 min async)

Minute-by-minute plan

  1. **00:00-10:00 Triage philosophy**
    • Open with: “You have 100 errors flagged in your volume and time to fix 20. Which 20 do you choose?”
    • Discuss: visual salience does not equal scientific importance. The ugliest error (a weird tentacle from a merge) may be less important than a subtle split in a key neuron.
    • Introduce impact-weighting framework.
    • Show a concrete example: two errors side by side, one visually dramatic but low-impact, one subtle but high-impact. Ask learners which they would fix first and why.
  2. **10:00-24:00 Queue classification exercise**
    • Present 12 pre-identified errors with brief descriptions. Learners work in pairs to:
      • Classify each by type (merge/split/boundary)
      • Estimate impact (high/medium/low) based on segment size and analysis relevance
      • Rank the top 5 for correction
    • Debrief: compare rankings across pairs. Where do teams disagree? Disagreements often reveal implicit assumptions about what matters.
  3. **24:00-38:00 Correction sprint**
    • Learners fix their top 5 errors in the practice dataset.
    • Instructor circulates: “Show me why you think this is a merge error.” “What evidence did you check before splitting?”
    • Emphasis on verification after each correction.
    • Common pitfall to watch for: learners who split a merge but forget to re-merge the orphaned fragment with the correct parent segment.
  4. **38:00-50:00 Threshold-based release decisions**
    • Compute metrics before and after the correction sprint.
    • Introduce release thresholds: “If ERL > 30 um and synapse F1 > 0.80, we release. If not, more proofreading.”
    • Group discussion: are we above threshold? If not, what would we fix next?
    • Key teaching moment: the threshold should be set before proofreading begins, not adjusted after seeing the results. Moving the goalposts undermines the purpose of having thresholds.
  5. **50:00-60:00 Competency check**
    • Each learner writes a 4-sentence “release recommendation memo”:
      • Current quality metrics
      • What was fixed
      • What remains unfixed and why
      • Go/no-go recommendation
    • Exit ticket: “One rule for when an error MUST be fixed before release.”

Studio activity: release decision simulation (60-75 minutes)

Scenario: You are the QC lead for a 100x100x100 um subvolume that will be used in a paper analyzing reciprocal connectivity between L2/3 pyramidal cells. The segmentation has been through one round of automated error detection. You need to decide: is this subvolume ready for analysis?

Task sequence:

  1. Review the automated error report: 45 flagged errors (18 merges, 20 splits, 7 uncertain).
  2. Triage: classify each by impact on the reciprocal connectivity analysis. Which errors could create false reciprocal connections? Which could hide real ones?
  3. Fix the top 15 errors, documenting each correction with a one-line rationale.
  4. Compute before/after metrics (provided metric computation script).
  5. Write a 1-page release recommendation memo with: metrics summary, corrections summary, remaining risks, and go/no-go recommendation with explicit reasoning.

Guidance for triage step: A merge error between two L2/3 pyramidal cells that are both pre- and post-synaptic to each other could create a spurious reciprocal connection or mask a real one – this is maximally damaging for the planned analysis. A split error on a glial process has zero impact on the reciprocal connectivity question. Errors on inhibitory interneurons fall in between: they don’t directly affect pyramidal-to-pyramidal reciprocal connections but could corrupt the broader circuit context. Learners should articulate this kind of reasoning for each error in their triage table.

Time management: Learners often spend too long on the triage step and run out of time for the correction sprint. Suggest a time budget: 20 minutes for triage, 30 minutes for corrections, 15 minutes for metrics and memo writing. The memo is the most important deliverable – a rushed memo undermines the entire exercise.

Expected outputs:

Assessment rubric

Content library references

Teaching resources

References

Quick practice prompt

Write one rule for when an error must be fixed before release. Your rule should specify: (1) the type of error, (2) the condition under which it is mandatory to fix, and (3) why that condition matters for downstream analysis. Example format: “A [type] error must be fixed before release when [condition], because [scientific reasoning].”

Teaching Materials

Activity Worksheet

Learner worksheet aligned to the studio activity and rubric.

Open worksheet

Slide Source

Marp source file for editing and rendering.

course/decks/marp/modules/module07.marp.md

Related Content