Skip to main content
MediaScribe
Deadline Extended: ADA compliance deadline moved to April 26, 2027.Learn what changed →

You are on lesson 3 of 11 in the course Path 1: Captions.

Module 1.1: What Makes Captions Compliant? Quality Standards

When someone asks if your captions are "compliant," they're really asking whether your captions serve the people who rely on them. The Web Content Accessibility Guidelines (WCAG) 2.1 Level AA requires captions for all video content with audio, but compliance goes beyond just having captions. Those captions need to meet specific quality standards.

Poor captions can frustrate viewers more than no captions at all. Imagine following a city council meeting where captions lag three seconds behind, miss critical votes, or block speakers' faces. A resident trying to follow along sees garbled text, misses the final vote count, and can't tell who's speaking. That's not accessibility. That's a barrier dressed up as accommodation.

The question isn't whether you have captions. The question is whether your captions actually work. This article breaks down the four quality standards that separate captions that help from captions that hinder.

The four pillars of caption quality

The Federal Communications Commission (FCC) established four quality standards for closed captions in 21 CFR Part 79. While written for broadcast television, these standards apply equally to web video and government recordings.


Understanding WCAG vs. FCC requirements

These two frameworks serve different purposes and apply to different contexts. WCAG 2.1 Level AA is a web accessibility standard — it defines what's required for digital content to be accessible, and it's what most government agencies must meet under Section 508 and ADA obligations. The FCC's caption quality rules (21 CFR Part 79) are broadcast regulations that apply to television programming.

However, the FCC's four quality pillars — accuracy, synchronization, completeness, and placement — are the most practical, widely recognized benchmarks for what "good captions" actually look like. That's why we use them as a quality framework throughout this course, even when your legal obligation flows from WCAG or Section 508 rather than FCC broadcast rules.

The key distinction: WCAG tells you that you need captions. The FCC framework tells you what good captions look like. Meeting FCC-level quality standards will satisfy your WCAG requirements, but FCC exemption criteria (such as those for certain low-viewership programming) do not exempt you from WCAG or ADA obligations for web-published content.


Accuracy: getting the words right

Accuracy means captions match spoken content word-for-word. The FCC requires 99% accuracy for pre-recorded broadcast content. For live captions, 95% accuracy is generally acceptable, though agencies may set higher standards based on content complexity.

That might sound like a small difference, but consider what it means in practice. A two-hour meeting with 15,000 spoken words allows 750 errors at 95% accuracy. That's one mistake every 15 seconds. For critical proceedings like budget votes or zoning decisions, even 95% might not be good enough.

Errors include substitutions ("their" vs. "there"), omissions, insertions, and formatting mistakes. Context matters. A misspelled name or "approved" transcribed as "a proof" has more impact than a filler word error. The former changes meaning. The latter just adds noise.

Common accuracy challenges:

  • Proper nouns (local street names, titles, community-specific terms)

  • Homonyms ("council" vs. "counsel")

  • Technical jargon and government acronyms

  • Cross-talk from multiple simultaneous speakers

The solution isn't just better technology. It's building custom vocabularies with local names, training systems with your specific terminology, and reviewing captions before archiving. Live captions will have errors. Archived captions shouldn't.

Synchronization: timing matters

FCC guidelines require captions to appear within three seconds of corresponding audio. For pre-recorded content, captions should align as closely as possible. For live content, consistent latency is better than erratic timing.

Timing affects comprehension. Poor synchronization forces viewers to choose between reading captions and watching visual action. Caption duration matters too. Captions should remain visible for at least one second and typically no more than six seconds, allowing comfortable reading at 160-180 words per minute.

Completeness: caption everything that matters

Program completeness means captions include all spoken dialogue and relevant non-speech information. This goes beyond transcribing words. Completeness requires capturing the full communication environment.

Required elements:

  • Speaker identification: Mark speaker changes using names or roles (>> MAYOR THOMPSON:). This is especially critical in government meetings where knowing who said what has legal significance. Without speaker labels, caption-only viewers lose track of debate flow and can't follow who supports or opposes motions.

  • Non-speech audio: Describe sounds that convey information [applause], [gavel strikes], [alarm sounding]. These cues provide context. Applause after a statement signals community support. A gavel strike marks an official action. An alarm interrupts proceedings. Caption-only viewers miss these unless you describe them.

  • Tone indicators: When tone changes meaning, reflect it through emphasis or description [sarcastically]. Sarcasm doesn't translate through text alone. Neither does emphasis that changes meaning from statement to question.

  • Music: Indicate when songs or music play [instrumental music playing]. If your meeting includes a presentation with background music or a ceremonial element with singing, note it.

Don't caption background noise that doesn't convey information. Distant traffic, general crowd murmur, and HVAC noise add nothing. Only caption sounds that provide context or meaning. The test is simple: would a sighted viewer gain information from this sound? If yes, caption it. If no, skip it.

Placement: don't block important information

Caption placement determines where text appears. Poor placement can obstruct faces, cover on-screen text, or hide visual elements. Best practices:

  • Default to bottom center for general discussion

  • Move captions when lower-third graphics or presentations occupy bottom area

  • Ensure adequate contrast (black on white or white on black)

  • Keep placement predictable so viewers know where to look

Compliant vs. non-compliant examples

Understanding quality standards becomes clearer with concrete examples. Here's what compliant and non-compliant captions look like in practice.

Budget vote non-compliant: "...and the motion is seconded we will now fold on resolution 2024-17..."

Problems: No speaker ID, transcription error ("fold" instead of "vote"), missing outcome.

Budget vote compliant: >> COUNCIL PRESIDENT MARTINEZ: And the motion is seconded. We will now vote on Resolution 2024-17. All in favor? [multiple voices: aye] Any opposed? [silence] The resolution passes unanimously.

Why it works: Clear speaker ID, accurate transcription, audio cues included, outcome stated.

Public comment non-compliant: "I've lived on main street for 30 years and this project will destroy our neighborhood we don't want it..."

Problems: No speaker ID, no capitalization, run-on sentence, incomplete statement.

Public comment compliant: >> RESIDENT JAMES CHEN: I've lived on Main Street for 30 years, and this project will destroy our neighborhood. We don't want it. The increased traffic will make our street unsafe for children. I'm asking the council to vote no.

Why it works: Speaker identified, proper punctuation, complete context.

Planning presentation non-compliant: "As you can see the setbacks meet code in all areas..."

Problems: No speaker ID, references visual without describing it, assumes viewer can see screen.

Planning presentation compliant: >> ARCHITECT SARAH WILLIAMS: This site plan shows setback measurements. The north property line has a 25-foot setback, the south has 30 feet, and both east and west sides exceed the required 20-foot minimum. All setbacks meet code requirements.

Why it works: Speaker identified, visual content described, measurements provided so caption-only viewers get same information as those who can see the slide.

Self-audit checklist

Evaluate your caption quality by reviewing at least three recent meetings. Rate each area as Strong, Adequate, or Needs Improvement. This systematic approach helps you identify patterns rather than one-off errors.

A practical two-page tool for government agencies to evaluate caption quality across the four FCC compliance pillars: accuracy, synchronization, completeness, and placement. Includes audit criteria, rating scales, and action planning sections to identify issues, assign responsibility, and track improvement over time. A PDF version of the checklist can be downloaded here.

Accuracy audit:

  • Proper nouns spelled correctly?

  • Technical terms transcribed accurately?

  • Numbers captured correctly?

  • Overall accuracy meets 95% standard?

Pick a two-minute segment from each meeting. Count every error (substitutions, omissions, insertions). Divide errors by total words. If you get more than 5% error rate, accuracy needs work.

Synchronization audit:

  • Captions appear within 3 seconds of audio?

  • Delay consistent throughout?

  • Display duration comfortable for reading?

Watch several segments with captions turned on. Do captions appear before speech finishes? That spoils content. Do they lag so far behind you're reading last sentence while speaker moved on? That creates confusion.

Completeness audit:

  • Speaker identification clear?

  • Non-speech audio described?

  • Visual references explained?

Turn off the video and listen to only captions read aloud. Can you follow the meeting? If not, you're missing completeness elements.

Placement audit:

  • Captions don't block faces or on-screen text?

  • Adequate contrast for easy reading?

  • Placement consistent and predictable?

Watch a meeting with presentations or name plates. Do captions cover the information? Watch someone giving public comment. Can you still see their face?

Action planning:

  • Identify your top three issues

  • Determine if each is a system, configuration, or workflow problem

  • Set specific improvement targets with timelines

  • Assign responsibility to specific staff

  • Schedule follow-up audit in 90 days

Most problems fall into categories. System problems need vendor help or equipment upgrades. Configuration problems need settings adjustments. Workflow problems need process changes. Knowing which type guides your solution.

Moving forward with confidence

Caption compliance is an ongoing commitment, not a one-time achievement. Technology helps enormously, but someone still needs to review captions periodically, maintain vocabulary lists, and ensure proper system configuration.

Most caption quality problems have straightforward solutions. Training your system with local terms fixes accuracy issues. Adjusting placement settings prevents visual obstruction. Building review processes catches problems before they become patterns. The best agencies treat caption quality as part of standard meeting workflow, not a separate accessibility project.

Remember why this matters. Compliant captions serve real people: residents who are deaf, employees who process information differently, reporters in noisy environments, people learning English. When your captions are accurate, synchronized, complete, and properly placed, you're not checking a box. You're making good on the promise of accessible government.


MediaScribe integration

MediaScribe helps government agencies maintain caption quality through built-in tools designed for the four compliance pillars. The system's custom vocabulary feature improves accuracy for local names and terms. Real-time synchronization maintains consistent timing for live captions. Speaker identification and non-speech audio notation ensure completeness. Multiple output formats support both in-room displays and broadcast channels with appropriate caption placement.

After meetings conclude, MediaScribe's caption editor allows staff to review and refine automated captions for archived recordings. The platform's quality control workflow makes it straightforward to conduct the self-audits described in this article.