Xenon Post
Xenon Post — Tools & Reference

Understanding Codecs
& Container Files

Codec and container are two words that are often used interchangeably but mean two entirely different things. Understanding the distinction — and knowing which formats appear at which stage of a production — is essential for anyone navigating a modern post pipeline.

01 — The container and the codec

When you look at a file on a drive — say, Interview_Cam-A_001.mov — what you are seeing is a container file. The .mov extension tells you the format of the wrapper. It tells you nothing about how the video and audio inside it are compressed.

The container is the file format itself: a structured wrapper that holds one or more streams of data. A video container typically holds a video stream, one or more audio streams, and often additional data such as timecode, subtitle tracks, and embedded metadata. Think of it as a box.

The codec — short for compressor/decompressor — is the algorithm that encodes and decodes the video data inside that box. It determines how images are compressed to reduce file size, and how they are reconstructed during playback. The same container can hold video compressed with entirely different codecs: a .mov file might contain ProRes 422, H.264, or DNxHD depending on what was put into it.

A useful analogy: the container is an envelope. The codec determines what language the letter inside is written in. You can receive envelopes of the same size and shape carrying letters in completely different languages — and you need the right decoder to read each one.

.MOV / .MP4 / .MXF — THE CONTAINERVIDEO STREAMProRes 422 HQ · 10-bit · 4:2:2AUDIO — CH 1/2PCM · 48kHz · 24-bitAUDIO — CH 3/4PCM · 48kHzTIMECODE25fps · Drop / NDMETADATACamera · Reel · Scene · TakeTHE SAME CONTAINERcan hold entirely different codecsfile.mov → ProRes 422 HQeditorial / post productionfile.mov → H.264 / AVCweb / client review deliveryfile.mov → DNxHD 115Avid Media Composer workflowcodecvariesContainer (.mov) stays the sameCodec (the compression) changes
Fig. 1 — Anatomy of a container file. Multiple streams live inside a single wrapper. The same container format can hold completely different codecs — the extension alone tells you nothing about the compression.
Container
The wrapper
Organises video, audio, timecode, and metadata into a single file. Determines file extension (.mov, .mp4, .mxf). Does not determine quality or compression.
Codec
The compression algorithm
Defines how the image data is encoded and decoded. Determines quality, file size, editability, and compatibility. Can be swapped independently of the container.

02 — How codecs compress video

All digital video compression involves discarding or approximating some information to reduce file size. The codec determines how much is discarded, what kind of information is prioritised, and how the original signal can be reconstructed. There are several axes on which codecs differ.

Intra-frame vs inter-frame compression

This is the most fundamental division between codec types and the one that most directly affects editing.

Intra-frame codecs — sometimes called I-frame-only codecs — compress each frame of video completely independently. Every frame contains all the information needed to reconstruct it without reference to any other frame. ProRes, DNxHR, and JPEG 2000 are all intra-frame. They produce larger files than inter-frame alternatives, but any frame can be read and decoded independently. This makes editing fast and accurate: the NLE can seek to any point in the timeline instantly.

Inter-frame codecs use temporal compression — they look across multiple frames and only encode the differences between them. H.264 and H.265 use three frame types: I-frames (complete reference frames), P-frames (which contain only the difference from the previous frame), and B-frames (which reference both the previous and next frames). This produces dramatically smaller files, but to decode any given frame the player may need to assemble data from several surrounding frames. Editing highly compressed inter-frame footage is computationally demanding and is why proxies are common in H.264/H.265 workflows.

The traditional rule of thumb: intra-frame codecs are for production and post, inter-frame codecs are for delivery. This remains broadly true — but inter-frame codecs have also found a valuable second role as proxy formats, particularly in RAW workflows.

H.264 and H.265 as proxy formats in RAW pipelines

In our day-to-day work at Xenon Post, H.264 and H.265 appear not as acquisition formats but as the output of a proxy generation step. The camera originals are RAW — ARRIRAW, R3D, BRAW — and it is from those files that lightweight proxy copies are generated for editing.

DaVinci Resolve’s built-in proxy generator, and Blackmagic’s standalone Proxy Generator application, can process RAW camera originals and output H.264 or H.265 proxy files at a fraction of the original file size. These proxies are timecode-matched and file-name-matched to their originals. An editor can cut on them on almost any machine — a laptop on location, a remote editor’s workstation — without needing access to the full RAW files at all. When the edit is complete, Resolve relinks the timeline to the RAW originals for grading and output, and the proxies are discarded.

H.264 and H.265 are well suited to this role precisely because of their efficiency: a proxy of a large RAW file can be small enough to share over the internet or store on a laptop drive, while still being visually accurate enough to cut on confidently. The inter-frame compression that makes them unsuitable as mastering formats is exactly what makes them practical as proxies.

Bit depth

Bit depth determines how many distinct tonal values each colour channel can represent. An 8-bit codec encodes 256 steps per channel — 256 levels of red, 256 of green, 256 of blue. This is adequate for web delivery and viewing, but provides limited headroom for grading: when you push shadows or highlights, you can reveal the steps between tonal values as visible banding.

10-bit encoding raises this to 1,024 steps per channel, which is the professional production standard and essential for any HDR workflow. 12-bit (4,096 steps) is used by higher-end cameras and VFX pipelines. OpenEXR images used in visual effects compositing typically operate at 16-bit or 32-bit float, preserving enormous tonal range for complex compositing operations.

Chroma subsampling

The human visual system is more sensitive to brightness (luminance) than to colour (chrominance). Chroma subsampling exploits this by recording colour at lower resolution than brightness information. The notation uses three numbers: the first represents the sampling block width (always 4), the second represents the number of colour samples in the first row, and the third in the second row.

Colour sample recordedInterpolated from neighbours4:4:4Full colour every pixel8 of 8 pixels sampledProRes 4444, RAW100% colour data4:2:2Colour every other column4 of 8 pixels sampledProRes 422, DNxHR, XAVC50% colour data4:2:0Every other column and row2 of 8 pixels sampledH.264, H.265, HEVC delivery25% colour dataLuma (brightness) is recorded at every pixel in all three schemes — only the chroma (colour) sample frequency changes
Fig. 2 — Chroma subsampling. Each cell represents one pixel. Amber dots indicate where colour is sampled; dim circles indicate where colour is interpolated from a neighbouring sample. Luma (brightness) is always sampled at every pixel regardless of scheme.

For green-screen work, keying, and any effects work requiring precise colour edges, 4:2:2 is the minimum practical requirement. The reduced chroma resolution of 4:2:0 makes clean key extraction significantly harder. For archive and highest-quality masters, 4:4:4 retains full colour at every pixel.

03 — Common container formats

Container formats are largely ecosystem decisions — the post house, the NLE, and the delivery platform each have preferences. The most important characteristic of a container is not quality (the codec determines that) but rather compatibility: what software can read and write it, and what delivery destinations accept it.

ContainerExtensionCommon UseNotes
QuickTime.movPost production, editorial, masteringApple's container. The dominant format in Mac-based post pipelines. Holds ProRes, H.264, DNxHD, and many others.
MPEG-4.mp4Delivery, streaming, web, client reviewBased on QuickTime's architecture. Universal compatibility. Almost always contains H.264 or H.265. The standard for any web-facing delivery.
MXF.mxfBroadcast delivery, archive, IMFMaterial eXchange Format. The SMPTE-defined professional broadcast standard. Required by many broadcasters. Holds DNxHD, XAVC, JPEG 2000, and others.
AVI.aviLegacy Windows workflowsMicrosoft's older container. Largely deprecated in professional post. Limited metadata support.
Matroska.mkvOpen-source, consumer distributionFlexible open container. Rarely required in professional delivery. Common for consumer high-quality video.
DPX.dpxVFX pipelines, film scanningSingle-frame container — one file per frame. The standard format for passing image sequences between VFX facilities and colourists.
OpenEXR.exrVFX compositing, HDR imageryILM's open format for high dynamic range images. Also single-frame. Standard in compositing pipelines. Supports multiple layers in a single file.
R3D.r3dRED camera acquisitionRED's proprietary format. Container and codec are inseparable. RAW sensor data with selectable REDCODE compression ratios.
BRAW.brawBlackmagic camera acquisitionBlackmagic RAW. Also inseparable container/codec. RAW sensor data with baked-in camera metadata and variable compression options.

04 — Common codecs by use case

Codecs divide naturally into three groups that reflect where in the pipeline they are used: acquisition codecs that cameras record to, editorial or mezzanine codecs used through post production, and delivery codecs for the final output. Using the wrong codec at the wrong stage — editing in a delivery codec, or delivering in an acquisition codec — is a very common and avoidable problem.

Acquisition & camera-native codecs

CodecTypeCamerasCharacteristics
ARRIRAWRAWARRI AlexaUncompressed or losslessly compressed sensor data. Maximum latitude, largest files. Requires debayering in post.
REDCODE / R3DRAWRED camerasProprietary wavelet-compressed RAW. Modern RED cameras offer three quality settings — HQ, MQ, and LQ — rather than fixed compression ratios. Still considered RAW — processing decisions deferred to post.
Blackmagic RAWRAWBlackmagic camerasRAW sensor data with baked-in camera metadata. Variable compression (Q0, Q5, 3:1, 5:1, 8:1, 12:1). Excellent performance in DaVinci Resolve.
X-OCNRAWSony Venice IIExtended tonal range Original Camera Negative. Sony's compressed RAW format for the Venice system. Available in three grades — XT (highest quality), ST (standard), and LT (light). Records up to 8.6K on the Venice II.
Cinema RAW LightRAWCanon EOS C300 III, C500 II, C70, R5 CCanon's compressed RAW acquisition format. Preserves the full dynamic range of the sensor with selectable compression. Readable natively in DaVinci Resolve.
XAVC / XAVC HSIntra / Long-GOPSony Venice II, FX9, FX6, FX3Sony's H.264-based (XAVC) and H.265-based (XAVC HS) codec family. Professional variants record 4:2:2 10-bit. Consumer XAVC S is 4:2:0.
ProRes (camera)IntraApple ProRes RAW cameras, Atomos recordersSome cameras record ProRes directly, eliminating the need for a post-transcode step. ProRes RAW is a separate (RAW-based) variant distinct from standard ProRes.
H.264 / AVCLong-GOPGH5, C70, various prosumerEfficient, universally readable, widely used in prosumer cameras. 4:2:0 in most camera implementations. Demanding to edit without transcode or proxy.
H.265 / HEVCLong-GOPSony, Canon, Panasonic, DJIApproximately 50% more efficient than H.264. Many 4K/60fps cameras use it. Even more demanding to edit than H.264 — proxy workflows strongly recommended.

Editorial & mezzanine codecs

Mezzanine codecs are high-quality, intra-frame codecs designed to be decoded quickly during editing, survive multiple transcode generations without significant quality loss, and serve as the master format through colour and finishing. They are not efficient enough for delivery, but they are the right tool for everything between acquisition and output.

CodecFamilyBit Depth / SamplingTypical Use
ProRes 4444 XQApple ProRes12-bit, 4:4:4Highest quality ProRes. VFX masters, motion graphics with alpha channel.
ProRes 4444Apple ProRes12-bit, 4:4:4, supports alphaVFX, compositing, wherever an alpha channel or maximum colour fidelity is needed.
ProRes 422 HQApple ProRes10-bit, 4:2:2High-quality mastering and delivery masters. Common output from DaVinci Resolve.
ProRes 422Apple ProRes10-bit, 4:2:2The everyday production workhorse. Online editorial, finishing.
ProRes 422 LTApple ProRes10-bit, 4:2:2Reduced bitrate. Useful when storage is limited but quality must be maintained.
ProRes 422 ProxyApple ProRes10-bit, 4:2:2Lowest bitrate ProRes. Offline editing and proxy workflows only — not for delivery.
DNxHR 444Avid DNxHR12-bit, 4:4:4Avid's highest quality. Equivalent to ProRes 4444 in Avid-based pipelines.
DNxHR HQXAvid DNxHR12-bit, 4:2:212-bit 4:2:2 variant. Sits between HQ and 444. Useful for HDR and high-end grading pipelines.
DNxHR HQAvid DNxHR8-bit, 4:2:2Avid's high-quality mezzanine. The Avid ecosystem equivalent to ProRes 422 HQ.
DNxHR LBAvid DNxHR8-bit, 4:2:2Low bandwidth. Offline/proxy use in Avid workflows.

ProRes and DNxHR are functionally equivalent in quality — they represent Apple’s and Avid’s respective approaches to the same problem. The right choice depends on your NLE and facility ecosystem, not on any inherent quality difference.

Delivery codecs

CodecContainerWhere RequiredNotes
H.264 / AVC.mp4 / .movWeb, Vimeo, YouTube, client delivery, broadcast OTTThe universal delivery codec. Highly efficient, universally readable. 4:2:0 in most delivery specs. Many broadcast platforms accept H.264 at 50–80 Mbps for HD delivery.
H.265 / HEVC.mp44K streaming, Netflix, Apple TV+, HDR delivery~50% smaller than H.264 at equivalent quality. Required by many platforms for 4K HDR delivery. Higher decode demands than H.264.
JPEG 2000.mxf (DCP)Theatrical DCPWavelet-based. Used exclusively in DCI Digital Cinema Packages. Intra-frame. Can be lossless or lossy within defined DCI bit rate limits.
IMF / JPEG 2000.mxf (IMF package)Netflix, Amazon, major streaming platformsInteroperable Master Format. A SMPTE-defined MXF package that allows platforms to create multiple versions (languages, HDR variants) from a single master.
DNxHD / DNxHR.mxfBroadcast delivery (some broadcasters)Some broadcasters specify MXF-wrapped DNxHD as their delivery format, particularly in Avid-centric facilities.

05 — RAW acquisition

RAW is frequently described as a codec, but it is more accurate to call it a data format. When a camera records RAW, it is capturing the unprocessed electrical signal from the sensor — the raw photon count at each photosite — before any in-camera processing has been applied. There is no white balance baked in, no sharpening, no noise reduction, and no colour science decisions locked to the image. All of that happens in post.

How camera sensors capture colour

Most digital cinema cameras use a single sensor with a colour filter array — typically a Bayer pattern — placed in front of the photosites. Each individual photosite can only record one colour (red, green, or blue) because of the filter above it. The camera records the intensity at each photosite, and a process called debayering (or demosaicing) reconstructs the full-colour image by interpolating the two missing colour values at each pixel from its neighbours. This is the step that converts RAW data into a viewable image, and it happens in post — not in the camera.

The consequence is that RAW images contain more information than a processed video codec, but they look flat and unprocessed until colour science is applied. The image that comes off a camera in RAW is not what the finished film looks like — it is the raw material from which the finished image is built.

Why shoot RAW

The primary advantage is exposure latitude. Because no in-camera processing decisions have been baked in, the full dynamic range captured by the sensor is preserved. Highlights and shadows that would be clipped in a processed codec remain recoverable in a RAW file, up to the physical limits of the sensor. A well-exposed RAW file from a modern cinema camera can yield 14 or more stops of usable dynamic range.

The secondary advantage is colour science flexibility. The choice of colour space, white balance interpretation, and tone mapping are all made in post — and can be changed non-destructively at any stage. Shooting RAW means a colourist is working with the maximum possible information, not a processed approximation of it.

The cost of RAW

RAW files are large. An ARRIRAW recording in open gate at 4K generates enormous data volumes — typical data rates range from 700 MB/s to over 2 GB/s for uncompressed or losslessly compressed formats. This has downstream effects: media costs are higher, data wrangling takes longer, and real-time playback requires either dedicated hardware or a transcode step. Most RAW workflows include a transcode to ProRes or DNxHR before editing, with the RAW originals retained as camera masters.

BRAW (Blackmagic RAW) occupies an interesting middle ground: it is a compressed RAW format that bakes in the camera’s colour metadata while still deferring the actual image development to post. It delivers much of the latitude and flexibility of traditional RAW at significantly lower data rates, and it plays back natively in DaVinci Resolve without a transcode step.

Planning a shoot and working out how much storage you need? We built a Data Rate Calculator that covers 45 cameras and every major acquisition codec — useful for estimating media costs and drives before you step on set.

06 — The post pipeline: from acquisition to delivery

Understanding individual codecs and containers is most useful in context — knowing how they connect through the stages of a production. A typical narrative or commercial pipeline moves through several distinct codec stages, each chosen for the demands of that stage rather than any single format serving all purposes.

ACQUISITIONPOSTDELIVERYCamera / RAWARRIRAW · R3D · BRAWOffload & VerifyChecksum · 2× backuptranscodeEditorial CodecProRes 422 · DNxHREdit + OnlinePremiere · Avid · FCPXGrade & FinishDaVinci ResolveoutputproxyofflineStreaming / WebH.264 · H.265 in .mp4BroadcastMXF · ProRes 422 HQTheatrical DCPJPEG 2000 in MXFCamera MastersArchive — do not editPrimary pipelineArchive / alternate path
Fig. 3 — A typical post production pipeline showing codec transitions from camera acquisition through editorial, grade, and delivery. Camera masters are archived separately and never used as edit media. Delivery splits to multiple output codecs simultaneously.

The mezzanine principle

A key concept the pipeline diagram illustrates is the mezzanine file — a high-quality intermediate that exists purely to serve the post process. Camera originals are archived and never touched again. An editorial transcode (ProRes 422 or DNxHR) is made from them, and it is this file that the editor works with day to day. From the finished grade, delivery files are transcoded downward for each platform. The mezzanine sits in the middle of this chain, absorbing the repeated read/write and decode operations that would degrade a more compressed format.

The proxy workflow

On projects where the full-resolution editorial codec is still too demanding for a given edit system — common with RAW 4K or 6K material, or on lighter laptops — a proxy is created alongside the full-resolution file. ProRes 422 Proxy or DNxHR LB are the standard choices: small enough to edit smoothly on almost any machine, but timecode-matched and file-name-matched to the high-resolution originals. When the edit is complete, the NLE reconforms the timeline to the full-resolution files for finishing. The proxy is discarded.

A common and costly mistake: delivering from the proxy rather than the original, or losing track of which files are proxies and which are camera masters. Rigorous file naming and folder structure at the data wrangling stage prevents this — and is one of the less glamorous but most valuable parts of what a good post facility manages.

07 — Quick reference

TermWhat it meansCommon confusion
ContainerThe file wrapper (.mov, .mp4, .mxf). Holds video, audio, timecode, metadata.File extension does not tell you the codec or quality of the content inside.
CodecThe compression algorithm applied to the video inside the container.ProRes and H.264 can both live inside a .mov file. Same wrapper, very different files.
Intra-frameEach frame compressed independently. ProRes, DNxHR, JPEG 2000.Larger files than inter-frame but essential for editing and grading.
Inter-frame (Long-GOP)Frames encoded relative to neighbours. H.264, H.265.Very efficient for delivery but not suitable as an editorial or mastering format.
4:2:2Colour sampled at half horizontal resolution. Professional production standard.Not the same as 4:2:0. Keying and colour work require 4:2:2 as minimum.
4:2:0Colour sampled at half horizontal and half vertical resolution.Common in camera acquisition (H.264/H.265) and delivery — not suitable for VFX plates or keying.
RAWUnprocessed sensor data. Not a codec — a data format requiring development in post.Often confused with Log. Log is a processed, gamma-encoded signal. RAW is pre-processed sensor data.
ProxyA low-resolution copy of editorial media used for offline editing.Proxies must never be used for delivery. They must be reconformed before finishing.
MezzanineA high-quality intermediate file used through post. Lives between camera originals and delivery.Not the delivery format and not the camera original — a working master for the post pipeline.
Xenon Post

The right format at every stage

Codec and container decisions made early in a production have consequences that run all the way through to delivery. At Xenon Post, we build pipeline specifications before a project starts — so there are no surprises when it comes time to finish.

Talk to the team