Skip to main content

Sync Offset Calculator

Practice and calculate sync offset with our free tool. Includes worked examples, visual aids, and learning resources. Free to use with no signup required.

Share this calculator

Formula

Offset (ms) = Frames x (1000 / FPS)

Where Frames is the number of frame offsets, FPS is the video frame rate, and the result is in milliseconds. Audio samples offset = (ms / 1000) x sample_rate. Net offset = video_delay - audio_delay + measured_offset.

Worked Examples

Example 1: Film Post-Production Sync Correction

Problem: A 24 fps film has audio arriving 3 frames early. The audio sample rate is 48000 Hz. Calculate the offset in milliseconds and samples.

Solution: ms per frame = 1000 / 24 = 41.667 ms\nOffset = 3 frames x 41.667 ms = 125 ms\nSamples = (125 / 1000) x 48000 = 6000 samples\nCorrection: delay audio by 125 ms (6000 samples)

Result: 3 frames = 125 ms = 6,000 audio samples delay needed

Example 2: Broadcast Stream Sync Analysis

Problem: A 29.97 fps broadcast has a measured video processing delay of 80 ms and audio processing delay of 15 ms. What is the net offset?

Solution: Net offset = video delay - audio delay\n= 80 ms - 15 ms = 65 ms\nAudio arrives 65 ms before its corresponding video frame\nAt 29.97 fps: 65 / 33.37 = 1.95 frames\nThis exceeds the EBU recommended 40 ms tolerance

Result: Net offset: 65 ms (audio leads) = 1.95 frames - exceeds broadcast tolerance

Frequently Asked Questions

What is audio-video sync offset and why does it matter?

Audio-video sync offset refers to the time difference between the audio track and the corresponding video frames in a media file or broadcast. Even small offsets can create a jarring viewing experience, particularly noticeable in dialogue scenes where lip movements do not match the spoken words. The human brain is remarkably sensitive to audio-visual timing discrepancies, with most viewers detecting offsets as small as 45 milliseconds. Professional broadcasting standards such as EBU R37 recommend keeping sync offset within plus or minus 40 milliseconds for acceptable quality. In post-production, sync issues can arise from processing latency, format conversions, or editing operations.

How do frame rates affect sync offset calculations?

Frame rate directly determines the temporal resolution of video and thus the granularity of sync adjustments. At 24 fps (film standard), each frame spans approximately 41.67 milliseconds, while at 30 fps each frame is 33.33 ms, and at 60 fps each frame is 16.67 ms. Higher frame rates allow finer sync adjustments because the minimum adjustment unit (one frame) represents a smaller time interval. When converting between frame rates, fractional frame offsets can occur, introducing sub-frame sync errors. This is particularly problematic in pulldown conversions between 24 fps film and 29.97 fps NTSC video, where the 3:2 pulldown pattern can create periodic sync drift.

What is the relationship between audio sample rate and sync precision?

Audio sample rate determines the finest time resolution available for sync adjustments on the audio side. At 48 kHz, each sample represents approximately 20.83 microseconds, giving extremely precise timing control. At 44.1 kHz (CD quality), each sample is about 22.68 microseconds. When aligning audio to video frame boundaries, the number of samples per frame may not be an integer, creating a small but accumulating rounding error. For example, at 48 kHz and 24 fps, there are exactly 2000 samples per frame, which aligns perfectly. But at 48 kHz and 29.97 fps, there are approximately 1601.6 samples per frame, requiring careful handling to prevent drift over long durations.

How do professionals detect and fix sync issues?

Professionals use several techniques to detect sync issues. The simplest is a clapperboard or slate, which provides a sharp visual and audio reference point for alignment. Digital tools include waveform displays overlaid on video timelines, dedicated sync analysis software, and test patterns with embedded audio tones. To fix sync issues, editors can slip the audio track relative to video in their timeline, apply sample-accurate delays, or use automatic sync detection algorithms that match audio waveforms to visual cues. In live broadcasting, dedicated hardware syncronizers and frame synchronizers continuously monitor and correct the relationship between audio and video signals.

What is the difference between fixed offset and drift?

A fixed offset is a constant time difference between audio and video that remains the same throughout the entire duration of the content. It can be corrected by simply shifting one track by a fixed amount. Drift, on the other hand, is a progressive change in the offset over time, where the audio and video gradually move further apart. Drift is more problematic because a simple shift will only fix the sync at one point in time. Drift correction requires either resampling the audio to match the video rate, conforming the video to match the audio, or applying variable time stretching. Identifying whether a sync problem is fixed or drifting is the first step in choosing the correct repair approach.

How does network latency affect live streaming sync?

In live streaming, audio and video are often encoded separately and may traverse different processing paths before reaching the viewer. Audio typically has lower latency because audio encoders process smaller data chunks faster than video encoders. This differential can cause audio to arrive and play before the corresponding video frame. Adaptive bitrate streaming protocols like HLS and DASH use timestamp-based synchronization to realign streams at the player, but varying network conditions can cause buffer fluctuations that introduce temporary sync errors. Content delivery networks may add variable latency to different stream components, and player-side buffering strategies can further affect perceived synchronization.

References