**YCbCr**, **Y′CbCr**, or **Y Pb/Cb Pr/Cr**, also written as **YC _{B}C_{R}** or

Y′CbCr color spaces are defined by a mathematical coordinate transformation from an associated RGB primaries and white point. If the underlying RGB color space is absolute, the Y′CbCr color space is an absolute color space as well; conversely, if the RGB space is ill-defined, so is Y′CbCr. The transformation is defined in equations 32, 33 in ITU-T H.273. Nevertheless that rule does not apply to P3-D65 primaries used by Netflix with BT.2020-NCL matrix, so that means matrix was not derived from primaries, but now Netflix allows BT.2020 primaries (since 2021).^{[1]} The same happens with JPEG: it has BT.601 matrix derived from System M primaries, yet the primaries of most images are BT.709.

Cathode ray tube displays are driven by red, green, and blue voltage signals, but these RGB signals are not efficient as a representation for storage and transmission, since they have a lot of redundancy.

YCbCr and Y′CbCr are a practical approximation to color processing and perceptual uniformity, where the primary colors corresponding roughly to red, green and blue are processed into perceptually meaningful information. By doing this, subsequent image/video processing, transmission and storage can do operations and introduce errors in perceptually meaningful ways. Y′CbCr is used to separate out a luma signal (Y′) that can be stored with high resolution or transmitted at high bandwidth, and two chroma components (C_{B} and C_{R}) that can be bandwidth-reduced, subsampled, compressed, or otherwise treated separately for improved system efficiency.

One practical example would be decreasing the bandwidth or resolution allocated to "color" compared to "black and white", since humans are more sensitive to the black-and-white information (see image example to the right). This is called chroma subsampling.

YCbCr is sometimes abbreviated to **YCC**.
Typically the terms Y′CbCr, YCbCr, YPbPr and YUV are used interchangeably, leading to some confusion. The main difference is that YPbPr is used with analog images and YCbCr with digital images, leading to different scaling values for U_{max} and V_{max} (in YCbCr both are ) when converting to/from YUV. Y′CbCr and YCbCr differ due to the values being gamma corrected or not.

The equations below give a better picture of the common principles and general differences between these formats.

Y′CbCr signals (prior to scaling and offsets to place the signals into digital form) are called YPbPr, and are created from the corresponding gamma-adjusted RGB (red, green and blue) source using three defined constants K_{R}, K_{G}, and K_{B} as follows:

where K_{R}, K_{G}, and K_{B} are ordinarily derived from the definition of the corresponding RGB space, and required to satisfy .

The equivalent matrix manipulation is often referred to as the "color matrix":

And its inverse:

Here, the prime (′) symbols mean gamma correction is being used; thus R′, G′ and B′ nominally range from 0 to 1, with 0 representing the minimum intensity (e.g., for display of the color black) and 1 the maximum (e.g., for display of the color white). The resulting luma (Y) value will then have a nominal range from 0 to 1, and the chroma (P_{B} and P_{R}) values will have a nominal range from -0.5 to +0.5. The reverse conversion process can be readily derived by inverting the above equations.

When representing the signals in digital form, the results are scaled and rounded, and offsets are typically added. For example, the scaling and offset applied to the Y′ component per specification (e.g. MPEG-2^{[2]}) results in the value of 16 for black and the value of 235 for white when using an 8-bit representation. The standard has 8-bit digitized versions of C_{B} and C_{R} scaled to a different range of 16 to 240. Consequently, rescaling by the fraction (235-16)/(240-16) = 219/224 is sometimes required when doing color matrixing or processing in YCbCr space, resulting in quantization distortions when the subsequent processing is not performed using higher bit depths.

The scaling that results in the use of a smaller range of digital values than what might appear to be desirable for representation of the nominal range of the input data allows for some "overshoot" and "undershoot" during processing without necessitating undesirable clipping. This "headroom" and "toeroom"^{[3]} can also be used for extension of the nominal color gamut, as specified by xvYCC.

The value 235 accommodates a maximum overshoot of (255 - 235) / (235 - 16) = 9.1%, which is slightly larger than the theoretical maximum overshoot (Gibbs' Phenomenon) of about 8.9% of the maximum (black-to-white) step. The toeroom is smaller, allowing only 16 / 219 = 7.3% overshoot, which is less than the theoretical maximum overshoot of 8.9%. In addition, because values 0 and 255 are reserved in HDMI, the room is actually slightly less.

Since the equations defining Y′CbCr are formed in a way that rotates the entire nominal RGB color cube and scales it to fit within a (larger) YCbCr color cube, there are some points within the Y′CbCr color cube that *cannot* be represented in the corresponding RGB domain (at least not within the nominal RGB range). This causes some difficulty in determining how to correctly interpret and display some Y′CbCr signals. These out-of-range Y′CbCr values are used by xvYCC to encode colors outside the BT.709 gamut.

The form of Y′CbCr that was defined for standard-definition television use in the ITU-R BT.601 (formerly CCIR 601) standard for use with digital component video is derived from the corresponding RGB space (ITU-R BT.470-6 System M primaries) as follows:

From the above constants and formulas, the following can be derived for ITU-R BT.601.

Analog YPbPr from analog R'G'B' is derived as follows:

Digital Y′CbCr (8 bits per sample) is derived from analog R'G'B' as follows:

or simply componentwise

The resultant signals range from 16 to 235 for Y′ (Cb and Cr range from 16 to 240); the values from 0 to 15 are called *footroom*, while the values from 236 to 255 are called *headroom*. The same quantisation ranges, different for Y and Cb, Cr also apply to BT.2020 and BT.709.

Alternatively, digital Y′CbCr can derived from digital R'dG'dB'd (8 bits per sample, each using the full range with zero representing black and 255 representing white) according to the following equations:

In the formula below, the scaling factors are multiplied by . This allows for the value 256 in the denominator, which can be calculated by a single bitshift.

If the R'd G'd B'd digital source includes footroom and headroom, the footroom offset 16 needs to be subtracted first from each signal, and a scale factor of needs to be included in the equations.

The inverse transform is:

The inverse transform without any roundings (using values coming directly from ITU-R BT.601 recommendation) is:

This form of Y′CbCr is used primarily for older standard-definition television systems, as it uses an RGB model that fits the phosphor emission characteristics of older CRTs.

Main article: BT.709 |

A different form of Y′CbCr is specified in the ITU-R BT.709 standard, primarily for HDTV use. The newer form is also used in some computer-display oriented applications, as sRGB (though the matrix used for sRGB form of YCbCr, sYCC, is still BT.601). In this case, the values of Kb and Kr differ, but the formulas for using them are the same. For ITU-R BT.709, the constants are:

This form of Y′CbCr is based on an RGB model that more closely fits the phosphor emission characteristics of newer CRTs and other modern display equipment.^{[citation needed]}
The conversion matrices for BT.709 are these:

The definitions of the R', G', and B' signals also differ between BT.709 and BT.601, and differ within BT.601 depending on the type of TV system in use (625-line as in PAL and SECAM or 525-line as in NTSC), and differ further in other specifications. In different designs there are differences in the definitions of the R, G, and B chromaticity coordinates, the reference white point, the supported gamut range, the exact gamma pre-compensation functions for deriving R', G' and B' from R, G, and B, and in the scaling and offsets to be applied during conversion from R'G'B' to Y′CbCr. So proper conversion of Y′CbCr from one form to the other is not just a matter of inverting one matrix and applying the other. In fact, when Y′CbCr is designed ideally, the values of K_{B} and K_{R} are derived from the precise specification of the RGB color primary signals, so that the luma (Y′) signal corresponds as closely as possible to a gamma-adjusted measurement of luminance (typically based on the CIE 1931 measurements of the response of the human visual system to color stimuli).^{[4]}

The ITU-R BT.2020 standard uses the same gamma function as BT.709. It defines:^{[5]}

- Non-constant luminance Y'CbCr, similar to the previous entries, except with different K
_{B}and K_{R}. - Constant luminance Y'cCbcCrc, a formulation where Y' is the gamma-codec version of the true luminance.
^{[5]}

For both, the coefficients derived from the primaries are:

For NCL, the definition is classical: Y' = 0.2627R' + 0.6780 G' + 0.0593 B'; Cb = (B' - Y') / 1.8814; Cr = (R' - Y') / 1.4746. The encoding conversion can, as usual, be written as a matrix.^{[5]} The decoding matrix for BT.2020-NCL is this with 14 decimal places:

The smaller values in the matrix are not rounded; they are precise values. For systems with limited precision (8 or 10 bit, for example) a lower precision of the above matrix could be used, for example, retaining only 6 digits after decimal point.^{[6]}

The CL version, YcCbcCrc, codes:^{[5]}

- Y'c = (0.2627R + 0.6780 G + 0.0593 B)'. This is the gamma function applied to the true luminance calculated from linear RGB.
- Cbc = (B'-Y'c) / (-2 Nb) if B' < Y'c; (B'-Y'c) / (2 Pb) otherwise. Nb and Pb are the theoretical minimum and maximum of (B'-Y'c) corresponding to the gamut. The rounded "practical" values are Pb = 0.7910, Nb = -0.9702. The full derivation can be found in the recommendation.
^{[5]} - Crc = (R'-Y'c) / (-2 Nr) if R' < Y'c; (R'-Y'c) / (2 Pr) otherwise. Again, Pr and Nr are theoretical limits. The rounded values are Pr 0.4969, Nr = -0.8591.

The CL function can be used when preservation of luminance is of primary importance (see: Chroma subsampling § Gamma luminance error), or when "there is an expectation of improved coding efficiency for delivery." The specification refers to Report ITU-R BT.2246 on this matter.^{[5]} BT.2246 states that CL has improved compression efficiency and luminance preservation, but NCL will be more familiar to a staff that has previously handled color mixing and other production practices in HDTV YCbCr.^{[7]}

BT.2020 does not define PQ and thus HDR, it is further defined in SMPTE ST 2084 and BT.2100. BT.2100 will introduce the use of IC_{T}C_{P}, a semi-perceptual colorspace derived from linear RGB with good hue linearity. It is "near-constant luminance".^{[8]}

The SMPTE 240M standard (used on the MUSE analog HD television system) defines YCC with these coefficients:

The coefficients are derived from SMPTE 170M primaries and white point, as used in 240M standard.

JFIF usage of JPEG supports a modified Rec. 601 Y′CbCr where Y′, C_{B} and C_{R} have the full 8-bit range of [0...255].^{[9]} Below are the conversion equations expressed to six decimal digits of precision. (For ideal equations, see ITU-T T.871.^{[10]})
Note that for the following formulae, the range of each input (R,G,B) is also the full 8-bit range of [0...255].

And back:

The above conversion is identical to sYCC when the input is given as sRGB, except that IEC 61966-2-1:1999/Amd1:2003 only gives four decimal digits.

JPEG also defines a "YCCK" format from Adobe for CMYK input. In this format, the "K" value is passed as-is, while CMY are used to derive YCbCr with the above matrix by assuming R=1-C, G=1-M, and B=1-Y. As a result, a similar set of subsampling techniques can be used.^{[11]}

These coefficients are not in use and were never in use.^{[12]}

H.273 also describes constant and non-constant luminance systems which are derived strictly from primaries and white point, so that situations like sRGB/BT.709 default primaries of JPEG that use BT.601 matrix (that is derived from BT.470-6 System M) do not happen.

Prior to the development of fast SIMD floating-point processors, most digital implementations of RGB → Y′UV used integer math, in particular fixed-point approximations. Approximation means that the precision of the used numbers (input data, output data and constant values) is limited, and thus a precision loss of typically about the last binary digit is accepted by whoever makes use of that option in typically a trade-off to improved computation speeds.

Y′ values are conventionally shifted and scaled to the range [16, 235] (referred to as studio swing or "TV levels") rather than using the full range of [0, 255] (referred to as full swing or "PC levels"). This practice was standardized in SMPTE-125M in order to accommodate signal overshoots ("ringing") due to filtering.^{[13]} U and V values, which may be positive or negative, are summed with 128 to make them always positive, giving a studio range of 16–240 for U and V. (These ranges are important in video editing and production, since using the wrong range will result either in an image with "clipped" blacks and whites, or a low-contrast image.)

These matrices round all factors to the closest 1/256 unit. As a result, only one 16-bit intermediate value is formed for each component, and a simple right-shift with rounding `(x + 128) >> 8`

can take care of the division.^{[13]}

For studio-swing:

For full-swing:

Google's Skia used to use the above 8-bit full-range matrix, resulting in a slight greening effect on JPEG images encoded by Android devices, more noticeable on repeated saving. The issue was fixed in 2016, when the more accurate version was used instead. Due to SIMD optimizations in libjpeg-turbo, the accurate version is actually faster.^{[14]}

RGB files are typically encoded in 8, 12, 16 or 24 bits per pixel. In these examples, we will assume 24 bits per pixel, which is written as RGB888. The standard byte format is simply `r0, g0, b0, r1, g1, b1, ...`

.

YCbCr Packed pixel formats are often referred to as "YUV". Such files can be encoded in 12, 16 or 24 bits per pixel. Depending on subsampling, the formats can largely be described as 4:4:4, 4:2:2, and 4:2:0p. The apostrophe after the Y is often omitted, as is the "p" (for planar) after YUV420p. In terms of actual file formats, 4:2:0 is the most common, as the data is more reduced, and the file extension is usually ".YUV". The relation between data rate and sampling (A:B:C) is defined by the ratio between Y to U and V channel.^{[15]}^{[16]} The notation of "YUV" followed by three numbers is vague: the three numbers could refer to the subsampling (as is done in "YUV420"), or it could refer to bit depth in each channel (as is done in "YUV565"). The unambiguous way to refer to these formats is via the FourCC code.^{[17]}

To convert from RGB to YUV or back, it is simplest to use RGB888 and 4:4:4. For 4:1:1, 4:2:2 and 4:2:0, the bytes need to be converted to 4:4:4 first.

4:4:4 is straightforward, as no pixel-grouping is done: the difference lies solely in how many bits each channel is given, and their arrangement. The basic `YUV3`

scheme uses 3 bytes per pixel, with the order `y0, u0, v0, y1, u1, v1`

(using "u" for Cb and "v" for Cr; the same applies to content below).^{[17]} In computers, it is more common to see a `AYUV`

format, which adds an alpha channel and goes `a0, y0, u0, v0, a1, y1, u1, v1`

, because groups of 32-bits are easier to deal with.^{[15]}

4:2:2 groups 2 pixels together horizontally in each conceptual "container". Two main arrangements are:^{[16]}

- YUY2: also called YUYV, runs in the format
`y0, u, y1, v`

. - UYVY: the byte-swapped reverse of YUY2, runs in the format
`u, y0, v, y1`

.

4:1:1 is rarely used. Pixels are in horizontal groups of 4.^{[16]}

4:2:0 is very commonly used. The main formats are IMC2, IMC4, YV12, and NV12.^{[16]} All of these four formats are "planar", meaning that the Y, U, and V values are grouped together instead of interspersed. They all occupy 12 bits per pixel, assuming a 8-bit channel.

- IMC2 first lays the full images out in Y. It then arranges each line of chroma in the order of V
_{0}... V_{n}, U_{0}... U_{n}, where*n*is the number of chroma samples per line, equal to half the width of Y. - IMC4 is similar to IMC2, except it runs in U
_{0}... U_{n}, V_{0}... V_{n}. - YV12 is a simpler design and is more commonly used. The entire image in Y is written out, followed by the image in V, then by the whole image in U.
- NV12 is possibly the most commonly-used 8-bit 4:2:0 format. It is the default for Android camera preview.
^{[18]}The entire image in Y is written out, followed by interleaved lines that go U_{0}, V_{0}, U_{1}, V_{1}, etc.

There are also "tiled" variants of planar formats.^{[19]}