<title>Image Formats</title> <para>The V4L2 API was primarily designed for devices exchanging image data with applications. The <structname>v4l2_pix_format</structname> and <structname>v4l2_pix_format_mplane </structname> structures define the format and layout of an image in memory. The former is used with the single-planar API, while the latter is used with the multi-planar version (see <xref linkend="planar-apis"/>). Image formats are negotiated with the &VIDIOC-S-FMT; ioctl. (The explanations here focus on video capturing and output, for overlay frame buffer formats see also &VIDIOC-G-FBUF;.)</para> <section> <title>Single-planar format structure</title> <table pgwide="1" frame="none" id="v4l2-pix-format"> <title>struct <structname>v4l2_pix_format</structname></title> <tgroup cols="3"> &cs-str; <tbody valign="top"> <row> <entry>__u32</entry> <entry><structfield>width</structfield></entry> <entry>Image width in pixels.</entry> </row> <row> <entry>__u32</entry> <entry><structfield>height</structfield></entry> <entry>Image height in pixels.</entry> </row> <row> <entry spanname="hspan">Applications set these fields to request an image size, drivers return the closest possible values. In case of planar formats the <structfield>width</structfield> and <structfield>height</structfield> applies to the largest plane. To avoid ambiguities drivers must return values rounded up to a multiple of the scale factor of any smaller planes. For example when the image format is YUV 4:2:0, <structfield>width</structfield> and <structfield>height</structfield> must be multiples of two.</entry> </row> <row> <entry>__u32</entry> <entry><structfield>pixelformat</structfield></entry> <entry>The pixel format or type of compression, set by the application. This is a little endian <link linkend="v4l2-fourcc">four character code</link>. V4L2 defines standard RGB formats in <xref linkend="rgb-formats" />, YUV formats in <xref linkend="yuv-formats" />, and reserved codes in <xref linkend="reserved-formats" /></entry> </row> <row> <entry>&v4l2-field;</entry> <entry><structfield>field</structfield></entry> <entry>Video images are typically interlaced. Applications can request to capture or output only the top or bottom field, or both fields interlaced or sequentially stored in one buffer or alternating in separate buffers. Drivers return the actual field order selected. For details see <xref linkend="field-order" />.</entry> </row> <row> <entry>__u32</entry> <entry><structfield>bytesperline</structfield></entry> <entry>Distance in bytes between the leftmost pixels in two adjacent lines.</entry> </row> <row> <entry spanname="hspan"><para>Both applications and drivers can set this field to request padding bytes at the end of each line. Drivers however may ignore the value requested by the application, returning <structfield>width</structfield> times bytes per pixel or a larger value required by the hardware. That implies applications can just set this field to zero to get a reasonable default.</para><para>Video hardware may access padding bytes, therefore they must reside in accessible memory. Consider cases where padding bytes after the last line of an image cross a system page boundary. Input devices may write padding bytes, the value is undefined. Output devices ignore the contents of padding bytes.</para><para>When the image format is planar the <structfield>bytesperline</structfield> value applies to the largest plane and is divided by the same factor as the <structfield>width</structfield> field for any smaller planes. For example the Cb and Cr planes of a YUV 4:2:0 image have half as many padding bytes following each line as the Y plane. To avoid ambiguities drivers must return a <structfield>bytesperline</structfield> value rounded up to a multiple of the scale factor.</para></entry> </row> <row> <entry>__u32</entry> <entry><structfield>sizeimage</structfield></entry> <entry>Size in bytes of the buffer to hold a complete image, set by the driver. Usually this is <structfield>bytesperline</structfield> times <structfield>height</structfield>. When the image consists of variable length compressed data this is the maximum number of bytes required to hold an image.</entry> </row> <row> <entry>&v4l2-colorspace;</entry> <entry><structfield>colorspace</structfield></entry> <entry>This information supplements the <structfield>pixelformat</structfield> and must be set by the driver, see <xref linkend="colorspaces" />.</entry> </row> <row> <entry>__u32</entry> <entry><structfield>priv</structfield></entry> <entry>Reserved for custom (driver defined) additional information about formats. When not used drivers and applications must set this field to zero.</entry> </row> </tbody> </tgroup> </table> </section> <section> <title>Multi-planar format structures</title> <para>The <structname>v4l2_plane_pix_format</structname> structures define size and layout for each of the planes in a multi-planar format. The <structname>v4l2_pix_format_mplane</structname> structure contains information common to all planes (such as image width and height) and an array of <structname>v4l2_plane_pix_format</structname> structures, describing all planes of that format.</para> <table pgwide="1" frame="none" id="v4l2-plane-pix-format"> <title>struct <structname>v4l2_plane_pix_format</structname></title> <tgroup cols="3"> &cs-str; <tbody valign="top"> <row> <entry>__u32</entry> <entry><structfield>sizeimage</structfield></entry> <entry>Maximum size in bytes required for image data in this plane. </entry> </row> <row> <entry>__u16</entry> <entry><structfield>bytesperline</structfield></entry> <entry>Distance in bytes between the leftmost pixels in two adjacent lines.</entry> </row> <row> <entry>__u16</entry> <entry><structfield>reserved[7]</structfield></entry> <entry>Reserved for future extensions. Should be zeroed by the application.</entry> </row> </tbody> </tgroup> </table> <table pgwide="1" frame="none" id="v4l2-pix-format-mplane"> <title>struct <structname>v4l2_pix_format_mplane</structname></title> <tgroup cols="3"> &cs-str; <tbody valign="top"> <row> <entry>__u32</entry> <entry><structfield>width</structfield></entry> <entry>Image width in pixels.</entry> </row> <row> <entry>__u32</entry> <entry><structfield>height</structfield></entry> <entry>Image height in pixels.</entry> </row> <row> <entry>__u32</entry> <entry><structfield>pixelformat</structfield></entry> <entry>The pixel format. Both single- and multi-planar four character codes can be used.</entry> </row> <row> <entry>&v4l2-field;</entry> <entry><structfield>field</structfield></entry> <entry>See &v4l2-pix-format;.</entry> </row> <row> <entry>&v4l2-colorspace;</entry> <entry><structfield>colorspace</structfield></entry> <entry>See &v4l2-pix-format;.</entry> </row> <row> <entry>&v4l2-plane-pix-format;</entry> <entry><structfield>plane_fmt[VIDEO_MAX_PLANES]</structfield></entry> <entry>An array of structures describing format of each plane this pixel format consists of. The number of valid entries in this array has to be put in the <structfield>num_planes</structfield> field.</entry> </row> <row> <entry>__u8</entry> <entry><structfield>num_planes</structfield></entry> <entry>Number of planes (i.e. separate memory buffers) for this format and the number of valid entries in the <structfield>plane_fmt</structfield> array.</entry> </row> <row> <entry>__u8</entry> <entry><structfield>reserved[11]</structfield></entry> <entry>Reserved for future extensions. Should be zeroed by the application.</entry> </row> </tbody> </tgroup> </table> </section> <section> <title>Standard Image Formats</title> <para>In order to exchange images between drivers and applications, it is necessary to have standard image data formats which both sides will interpret the same way. V4L2 includes several such formats, and this section is intended to be an unambiguous specification of the standard image data formats in V4L2.</para> <para>V4L2 drivers are not limited to these formats, however. Driver-specific formats are possible. In that case the application may depend on a codec to convert images to one of the standard formats when needed. But the data can still be stored and retrieved in the proprietary format. For example, a device may support a proprietary compressed format. Applications can still capture and save the data in the compressed format, saving much disk space, and later use a codec to convert the images to the X Windows screen format when the video is to be displayed.</para> <para>Even so, ultimately, some standard formats are needed, so the V4L2 specification would not be complete without well-defined standard formats.</para> <para>The V4L2 standard formats are mainly uncompressed formats. The pixels are always arranged in memory from left to right, and from top to bottom. The first byte of data in the image buffer is always for the leftmost pixel of the topmost row. Following that is the pixel immediately to its right, and so on until the end of the top row of pixels. Following the rightmost pixel of the row there may be zero or more bytes of padding to guarantee that each row of pixel data has a certain alignment. Following the pad bytes, if any, is data for the leftmost pixel of the second row from the top, and so on. The last row has just as many pad bytes after it as the other rows.</para> <para>In V4L2 each format has an identifier which looks like <constant>PIX_FMT_XXX</constant>, defined in the <link linkend="videodev">videodev.h</link> header file. These identifiers represent <link linkend="v4l2-fourcc">four character (FourCC) codes</link> which are also listed below, however they are not the same as those used in the Windows world.</para> <para>For some formats, data is stored in separate, discontiguous memory buffers. Those formats are identified by a separate set of FourCC codes and are referred to as "multi-planar formats". For example, a YUV422 frame is normally stored in one memory buffer, but it can also be placed in two or three separate buffers, with Y component in one buffer and CbCr components in another in the 2-planar version or with each component in its own buffer in the 3-planar case. Those sub-buffers are referred to as "planes".</para> </section> <section id="colorspaces"> <title>Colorspaces</title> <para>[intro]</para> <!-- See proposal by Billy Biggs, video4linux-list@redhat.com on 11 Oct 2002, subject: "Re: [V4L] Re: v4l2 api", and http://vektor.theorem.ca/graphics/ycbcr/ and http://www.poynton.com/notes/colour_and_gamma/ColorFAQ.html --> <para> <variablelist> <varlistentry> <term>Gamma Correction</term> <listitem> <para>[to do]</para> <para>E'<subscript>R</subscript> = f(R)</para> <para>E'<subscript>G</subscript> = f(G)</para> <para>E'<subscript>B</subscript> = f(B)</para> </listitem> </varlistentry> <varlistentry> <term>Construction of luminance and color-difference signals</term> <listitem> <para>[to do]</para> <para>E'<subscript>Y</subscript> = Coeff<subscript>R</subscript> E'<subscript>R</subscript> + Coeff<subscript>G</subscript> E'<subscript>G</subscript> + Coeff<subscript>B</subscript> E'<subscript>B</subscript></para> <para>(E'<subscript>R</subscript> - E'<subscript>Y</subscript>) = E'<subscript>R</subscript> - Coeff<subscript>R</subscript> E'<subscript>R</subscript> - Coeff<subscript>G</subscript> E'<subscript>G</subscript> - Coeff<subscript>B</subscript> E'<subscript>B</subscript></para> <para>(E'<subscript>B</subscript> - E'<subscript>Y</subscript>) = E'<subscript>B</subscript> - Coeff<subscript>R</subscript> E'<subscript>R</subscript> - Coeff<subscript>G</subscript> E'<subscript>G</subscript> - Coeff<subscript>B</subscript> E'<subscript>B</subscript></para> </listitem> </varlistentry> <varlistentry> <term>Re-normalized color-difference signals</term> <listitem> <para>The color-difference signals are scaled back to unity range [-0.5;+0.5]:</para> <para>K<subscript>B</subscript> = 0.5 / (1 - Coeff<subscript>B</subscript>)</para> <para>K<subscript>R</subscript> = 0.5 / (1 - Coeff<subscript>R</subscript>)</para> <para>P<subscript>B</subscript> = K<subscript>B</subscript> (E'<subscript>B</subscript> - E'<subscript>Y</subscript>) = 0.5 (Coeff<subscript>R</subscript> / Coeff<subscript>B</subscript>) E'<subscript>R</subscript> + 0.5 (Coeff<subscript>G</subscript> / Coeff<subscript>B</subscript>) E'<subscript>G</subscript> + 0.5 E'<subscript>B</subscript></para> <para>P<subscript>R</subscript> = K<subscript>R</subscript> (E'<subscript>R</subscript> - E'<subscript>Y</subscript>) = 0.5 E'<subscript>R</subscript> + 0.5 (Coeff<subscript>G</subscript> / Coeff<subscript>R</subscript>) E'<subscript>G</subscript> + 0.5 (Coeff<subscript>B</subscript> / Coeff<subscript>R</subscript>) E'<subscript>B</subscript></para> </listitem> </varlistentry> <varlistentry> <term>Quantization</term> <listitem> <para>[to do]</para> <para>Y' = (Lum. Levels - 1) · E'<subscript>Y</subscript> + Lum. Offset</para> <para>C<subscript>B</subscript> = (Chrom. Levels - 1) · P<subscript>B</subscript> + Chrom. Offset</para> <para>C<subscript>R</subscript> = (Chrom. Levels - 1) · P<subscript>R</subscript> + Chrom. Offset</para> <para>Rounding to the nearest integer and clamping to the range [0;255] finally yields the digital color components Y'CbCr stored in YUV images.</para> </listitem> </varlistentry> </variablelist> </para> <example> <title>ITU-R Rec. BT.601 color conversion</title> <para>Forward Transformation</para> <programlisting> int ER, EG, EB; /* gamma corrected RGB input [0;255] */ int Y1, Cb, Cr; /* output [0;255] */ double r, g, b; /* temporaries */ double y1, pb, pr; int clamp (double x) { int r = x; /* round to nearest */ if (r < 0) return 0; else if (r > 255) return 255; else return r; } r = ER / 255.0; g = EG / 255.0; b = EB / 255.0; y1 = 0.299 * r + 0.587 * g + 0.114 * b; pb = -0.169 * r - 0.331 * g + 0.5 * b; pr = 0.5 * r - 0.419 * g - 0.081 * b; Y1 = clamp (219 * y1 + 16); Cb = clamp (224 * pb + 128); Cr = clamp (224 * pr + 128); /* or shorter */ y1 = 0.299 * ER + 0.587 * EG + 0.114 * EB; Y1 = clamp ( (219 / 255.0) * y1 + 16); Cb = clamp (((224 / 255.0) / (2 - 2 * 0.114)) * (EB - y1) + 128); Cr = clamp (((224 / 255.0) / (2 - 2 * 0.299)) * (ER - y1) + 128); </programlisting> <para>Inverse Transformation</para> <programlisting> int Y1, Cb, Cr; /* gamma pre-corrected input [0;255] */ int ER, EG, EB; /* output [0;255] */ double r, g, b; /* temporaries */ double y1, pb, pr; int clamp (double x) { int r = x; /* round to nearest */ if (r < 0) return 0; else if (r > 255) return 255; else return r; } y1 = (255 / 219.0) * (Y1 - 16); pb = (255 / 224.0) * (Cb - 128); pr = (255 / 224.0) * (Cr - 128); r = 1.0 * y1 + 0 * pb + 1.402 * pr; g = 1.0 * y1 - 0.344 * pb - 0.714 * pr; b = 1.0 * y1 + 1.772 * pb + 0 * pr; ER = clamp (r * 255); /* [ok? one should prob. limit y1,pb,pr] */ EG = clamp (g * 255); EB = clamp (b * 255); </programlisting> </example> <table pgwide="1" id="v4l2-colorspace" orient="land"> <title>enum v4l2_colorspace</title> <tgroup cols="11" align="center"> <colspec align="left" /> <colspec align="center" /> <colspec align="left" /> <colspec colname="cr" /> <colspec colname="cg" /> <colspec colname="cb" /> <colspec colname="wp" /> <colspec colname="gc" /> <colspec colname="lum" /> <colspec colname="qy" /> <colspec colname="qc" /> <spanspec namest="cr" nameend="cb" spanname="chrom" /> <spanspec namest="qy" nameend="qc" spanname="quant" /> <spanspec namest="lum" nameend="qc" spanname="spam" /> <thead> <row> <entry morerows="1">Identifier</entry> <entry morerows="1">Value</entry> <entry morerows="1">Description</entry> <entry spanname="chrom">Chromaticities<footnote> <para>The coordinates of the color primaries are given in the CIE system (1931)</para> </footnote></entry> <entry morerows="1">White Point</entry> <entry morerows="1">Gamma Correction</entry> <entry morerows="1">Luminance E'<subscript>Y</subscript></entry> <entry spanname="quant">Quantization</entry> </row> <row> <entry>Red</entry> <entry>Green</entry> <entry>Blue</entry> <entry>Y'</entry> <entry>Cb, Cr</entry> </row> </thead> <tbody valign="top"> <row> <entry><constant>V4L2_COLORSPACE_SMPTE170M</constant></entry> <entry>1</entry> <entry>NTSC/PAL according to <xref linkend="smpte170m" />, <xref linkend="itu601" /></entry> <entry>x = 0.630, y = 0.340</entry> <entry>x = 0.310, y = 0.595</entry> <entry>x = 0.155, y = 0.070</entry> <entry>x = 0.3127, y = 0.3290, Illuminant D<subscript>65</subscript></entry> <entry>E' = 4.5 I for I ≤0.018, 1.099 I<superscript>0.45</superscript> - 0.099 for 0.018 < I</entry> <entry>0.299 E'<subscript>R</subscript> + 0.587 E'<subscript>G</subscript> + 0.114 E'<subscript>B</subscript></entry> <entry>219 E'<subscript>Y</subscript> + 16</entry> <entry>224 P<subscript>B,R</subscript> + 128</entry> </row> <row> <entry><constant>V4L2_COLORSPACE_SMPTE240M</constant></entry> <entry>2</entry> <entry>1125-Line (US) HDTV, see <xref linkend="smpte240m" /></entry> <entry>x = 0.630, y = 0.340</entry> <entry>x = 0.310, y = 0.595</entry> <entry>x = 0.155, y = 0.070</entry> <entry>x = 0.3127, y = 0.3290, Illuminant D<subscript>65</subscript></entry> <entry>E' = 4 I for I ≤0.0228, 1.1115 I<superscript>0.45</superscript> - 0.1115 for 0.0228 < I</entry> <entry>0.212 E'<subscript>R</subscript> + 0.701 E'<subscript>G</subscript> + 0.087 E'<subscript>B</subscript></entry> <entry>219 E'<subscript>Y</subscript> + 16</entry> <entry>224 P<subscript>B,R</subscript> + 128</entry> </row> <row> <entry><constant>V4L2_COLORSPACE_REC709</constant></entry> <entry>3</entry> <entry>HDTV and modern devices, see <xref linkend="itu709" /></entry> <entry>x = 0.640, y = 0.330</entry> <entry>x = 0.300, y = 0.600</entry> <entry>x = 0.150, y = 0.060</entry> <entry>x = 0.3127, y = 0.3290, Illuminant D<subscript>65</subscript></entry> <entry>E' = 4.5 I for I ≤0.018, 1.099 I<superscript>0.45</superscript> - 0.099 for 0.018 < I</entry> <entry>0.2125 E'<subscript>R</subscript> + 0.7154 E'<subscript>G</subscript> + 0.0721 E'<subscript>B</subscript></entry> <entry>219 E'<subscript>Y</subscript> + 16</entry> <entry>224 P<subscript>B,R</subscript> + 128</entry> </row> <row> <entry><constant>V4L2_COLORSPACE_BT878</constant></entry> <entry>4</entry> <entry>Broken Bt878 extents<footnote> <para>The ubiquitous Bt878 video capture chip quantizes E'<subscript>Y</subscript> to 238 levels, yielding a range of Y' = 16 … 253, unlike Rec. 601 Y' = 16 … 235. This is not a typo in the Bt878 documentation, it has been implemented in silicon. The chroma extents are unclear.</para> </footnote>, <xref linkend="itu601" /></entry> <entry>?</entry> <entry>?</entry> <entry>?</entry> <entry>?</entry> <entry>?</entry> <entry>0.299 E'<subscript>R</subscript> + 0.587 E'<subscript>G</subscript> + 0.114 E'<subscript>B</subscript></entry> <entry><emphasis>237</emphasis> E'<subscript>Y</subscript> + 16</entry> <entry>224 P<subscript>B,R</subscript> + 128 (probably)</entry> </row> <row> <entry><constant>V4L2_COLORSPACE_470_SYSTEM_M</constant></entry> <entry>5</entry> <entry>M/NTSC<footnote> <para>No identifier exists for M/PAL which uses the chromaticities of M/NTSC, the remaining parameters are equal to B and G/PAL.</para> </footnote> according to <xref linkend="itu470" />, <xref linkend="itu601" /></entry> <entry>x = 0.67, y = 0.33</entry> <entry>x = 0.21, y = 0.71</entry> <entry>x = 0.14, y = 0.08</entry> <entry>x = 0.310, y = 0.316, Illuminant C</entry> <entry>?</entry> <entry>0.299 E'<subscript>R</subscript> + 0.587 E'<subscript>G</subscript> + 0.114 E'<subscript>B</subscript></entry> <entry>219 E'<subscript>Y</subscript> + 16</entry> <entry>224 P<subscript>B,R</subscript> + 128</entry> </row> <row> <entry><constant>V4L2_COLORSPACE_470_SYSTEM_BG</constant></entry> <entry>6</entry> <entry>625-line PAL and SECAM systems according to <xref linkend="itu470" />, <xref linkend="itu601" /></entry> <entry>x = 0.64, y = 0.33</entry> <entry>x = 0.29, y = 0.60</entry> <entry>x = 0.15, y = 0.06</entry> <entry>x = 0.313, y = 0.329, Illuminant D<subscript>65</subscript></entry> <entry>?</entry> <entry>0.299 E'<subscript>R</subscript> + 0.587 E'<subscript>G</subscript> + 0.114 E'<subscript>B</subscript></entry> <entry>219 E'<subscript>Y</subscript> + 16</entry> <entry>224 P<subscript>B,R</subscript> + 128</entry> </row> <row> <entry><constant>V4L2_COLORSPACE_JPEG</constant></entry> <entry>7</entry> <entry>JPEG Y'CbCr, see <xref linkend="jfif" />, <xref linkend="itu601" /></entry> <entry>?</entry> <entry>?</entry> <entry>?</entry> <entry>?</entry> <entry>?</entry> <entry>0.299 E'<subscript>R</subscript> + 0.587 E'<subscript>G</subscript> + 0.114 E'<subscript>B</subscript></entry> <entry>256 E'<subscript>Y</subscript> + 16<footnote> <para>Note JFIF quantizes Y'P<subscript>B</subscript>P<subscript>R</subscript> in range [0;+1] and [-0.5;+0.5] to <emphasis>257</emphasis> levels, however Y'CbCr signals are still clamped to [0;255].</para> </footnote></entry> <entry>256 P<subscript>B,R</subscript> + 128</entry> </row> <row> <entry><constant>V4L2_COLORSPACE_SRGB</constant></entry> <entry>8</entry> <entry>[?]</entry> <entry>x = 0.640, y = 0.330</entry> <entry>x = 0.300, y = 0.600</entry> <entry>x = 0.150, y = 0.060</entry> <entry>x = 0.3127, y = 0.3290, Illuminant D<subscript>65</subscript></entry> <entry>E' = 4.5 I for I ≤0.018, 1.099 I<superscript>0.45</superscript> - 0.099 for 0.018 < I</entry> <entry spanname="spam">n/a</entry> </row> </tbody> </tgroup> </table> </section> <section id="pixfmt-indexed"> <title>Indexed Format</title> <para>In this format each pixel is represented by an 8 bit index into a 256 entry ARGB palette. It is intended for <link linkend="osd">Video Output Overlays</link> only. There are no ioctls to access the palette, this must be done with ioctls of the Linux framebuffer API.</para> <table pgwide="0" frame="none"> <title>Indexed Image Format</title> <tgroup cols="37" align="center"> <colspec colname="id" align="left" /> <colspec colname="fourcc" /> <colspec colname="bit" /> <colspec colnum="4" colname="b07" align="center" /> <colspec colnum="5" colname="b06" align="center" /> <colspec colnum="6" colname="b05" align="center" /> <colspec colnum="7" colname="b04" align="center" /> <colspec colnum="8" colname="b03" align="center" /> <colspec colnum="9" colname="b02" align="center" /> <colspec colnum="10" colname="b01" align="center" /> <colspec colnum="11" colname="b00" align="center" /> <spanspec namest="b07" nameend="b00" spanname="b0" /> <spanspec namest="b17" nameend="b10" spanname="b1" /> <spanspec namest="b27" nameend="b20" spanname="b2" /> <spanspec namest="b37" nameend="b30" spanname="b3" /> <thead> <row> <entry>Identifier</entry> <entry>Code</entry> <entry> </entry> <entry spanname="b0">Byte 0</entry> </row> <row> <entry> </entry> <entry> </entry> <entry>Bit</entry> <entry>7</entry> <entry>6</entry> <entry>5</entry> <entry>4</entry> <entry>3</entry> <entry>2</entry> <entry>1</entry> <entry>0</entry> </row> </thead> <tbody valign="top"> <row id="V4L2-PIX-FMT-PAL8"> <entry><constant>V4L2_PIX_FMT_PAL8</constant></entry> <entry>'PAL8'</entry> <entry></entry> <entry>i<subscript>7</subscript></entry> <entry>i<subscript>6</subscript></entry> <entry>i<subscript>5</subscript></entry> <entry>i<subscript>4</subscript></entry> <entry>i<subscript>3</subscript></entry> <entry>i<subscript>2</subscript></entry> <entry>i<subscript>1</subscript></entry> <entry>i<subscript>0</subscript></entry> </row> </tbody> </tgroup> </table> </section> <section id="pixfmt-rgb"> <title>RGB Formats</title> &sub-packed-rgb; &sub-sbggr8; &sub-sgbrg8; &sub-sgrbg8; &sub-srggb8; &sub-sbggr16; &sub-srggb10; &sub-srggb10alaw8; &sub-srggb10dpcm8; &sub-srggb12; </section> <section id="yuv-formats"> <title>YUV Formats</title> <para>YUV is the format native to TV broadcast and composite video signals. It separates the brightness information (Y) from the color information (U and V or Cb and Cr). The color information consists of red and blue <emphasis>color difference</emphasis> signals, this way the green component can be reconstructed by subtracting from the brightness component. See <xref linkend="colorspaces" /> for conversion examples. YUV was chosen because early television would only transmit brightness information. To add color in a way compatible with existing receivers a new signal carrier was added to transmit the color difference signals. Secondary in the YUV format the U and V components usually have lower resolution than the Y component. This is an analog video compression technique taking advantage of a property of the human visual system, being more sensitive to brightness information.</para> &sub-packed-yuv; &sub-grey; &sub-y10; &sub-y12; &sub-y10b; &sub-y16; &sub-uv8; &sub-yuyv; &sub-uyvy; &sub-yvyu; &sub-vyuy; &sub-y41p; &sub-yuv420; &sub-yuv420m; &sub-yvu420m; &sub-yuv410; &sub-yuv422p; &sub-yuv411p; &sub-nv12; &sub-nv12m; &sub-nv12mt; &sub-nv16; &sub-nv24; &sub-m420; </section> <section> <title>Compressed Formats</title> <table pgwide="1" frame="none" id="compressed-formats"> <title>Compressed Image Formats</title> <tgroup cols="3" align="left"> &cs-def; <thead> <row> <entry>Identifier</entry> <entry>Code</entry> <entry>Details</entry> </row> </thead> <tbody valign="top"> <row id="V4L2-PIX-FMT-JPEG"> <entry><constant>V4L2_PIX_FMT_JPEG</constant></entry> <entry>'JPEG'</entry> <entry>TBD. See also &VIDIOC-G-JPEGCOMP;, &VIDIOC-S-JPEGCOMP;.</entry> </row> <row id="V4L2-PIX-FMT-MPEG"> <entry><constant>V4L2_PIX_FMT_MPEG</constant></entry> <entry>'MPEG'</entry> <entry>MPEG multiplexed stream. The actual format is determined by extended control <constant>V4L2_CID_MPEG_STREAM_TYPE</constant>, see <xref linkend="mpeg-control-id" />.</entry> </row> <row id="V4L2-PIX-FMT-H264"> <entry><constant>V4L2_PIX_FMT_H264</constant></entry> <entry>'H264'</entry> <entry>H264 video elementary stream with start codes.</entry> </row> <row id="V4L2-PIX-FMT-H264-NO-SC"> <entry><constant>V4L2_PIX_FMT_H264_NO_SC</constant></entry> <entry>'AVC1'</entry> <entry>H264 video elementary stream without start codes.</entry> </row> <row id="V4L2-PIX-FMT-H264-MVC"> <entry><constant>V4L2_PIX_FMT_H264_MVC</constant></entry> <entry>'MVC'</entry> <entry>H264 MVC video elementary stream.</entry> </row> <row id="V4L2-PIX-FMT-H263"> <entry><constant>V4L2_PIX_FMT_H263</constant></entry> <entry>'H263'</entry> <entry>H263 video elementary stream.</entry> </row> <row id="V4L2-PIX-FMT-MPEG1"> <entry><constant>V4L2_PIX_FMT_MPEG1</constant></entry> <entry>'MPG1'</entry> <entry>MPEG1 video elementary stream.</entry> </row> <row id="V4L2-PIX-FMT-MPEG2"> <entry><constant>V4L2_PIX_FMT_MPEG2</constant></entry> <entry>'MPG2'</entry> <entry>MPEG2 video elementary stream.</entry> </row> <row id="V4L2-PIX-FMT-MPEG4"> <entry><constant>V4L2_PIX_FMT_MPEG4</constant></entry> <entry>'MPG4'</entry> <entry>MPEG4 video elementary stream.</entry> </row> <row id="V4L2-PIX-FMT-XVID"> <entry><constant>V4L2_PIX_FMT_XVID</constant></entry> <entry>'XVID'</entry> <entry>Xvid video elementary stream.</entry> </row> <row id="V4L2-PIX-FMT-VC1-ANNEX-G"> <entry><constant>V4L2_PIX_FMT_VC1_ANNEX_G</constant></entry> <entry>'VC1G'</entry> <entry>VC1, SMPTE 421M Annex G compliant stream.</entry> </row> <row id="V4L2-PIX-FMT-VC1-ANNEX-L"> <entry><constant>V4L2_PIX_FMT_VC1_ANNEX_L</constant></entry> <entry>'VC1L'</entry> <entry>VC1, SMPTE 421M Annex L compliant stream.</entry> </row> <row id="V4L2-PIX-FMT-VP8"> <entry><constant>V4L2_PIX_FMT_VP8</constant></entry> <entry>'VP8'</entry> <entry>VP8 video elementary stream.</entry> </row> </tbody> </tgroup> </table> </section> <section id="pixfmt-reserved"> <title>Reserved Format Identifiers</title> <para>These formats are not defined by this specification, they are just listed for reference and to avoid naming conflicts. If you want to register your own format, send an e-mail to the linux-media mailing list &v4l-ml; for inclusion in the <filename>videodev2.h</filename> file. If you want to share your format with other developers add a link to your documentation and send a copy to the linux-media mailing list for inclusion in this section. If you think your format should be listed in a standard format section please make a proposal on the linux-media mailing list.</para> <table pgwide="1" frame="none" id="reserved-formats"> <title>Reserved Image Formats</title> <tgroup cols="3" align="left"> &cs-def; <thead> <row> <entry>Identifier</entry> <entry>Code</entry> <entry>Details</entry> </row> </thead> <tbody valign="top"> <row id="V4L2-PIX-FMT-DV"> <entry><constant>V4L2_PIX_FMT_DV</constant></entry> <entry>'dvsd'</entry> <entry>unknown</entry> </row> <row id="V4L2-PIX-FMT-ET61X251"> <entry><constant>V4L2_PIX_FMT_ET61X251</constant></entry> <entry>'E625'</entry> <entry>Compressed format of the ET61X251 driver.</entry> </row> <row id="V4L2-PIX-FMT-HI240"> <entry><constant>V4L2_PIX_FMT_HI240</constant></entry> <entry>'HI24'</entry> <entry><para>8 bit RGB format used by the BTTV driver.</para></entry> </row> <row id="V4L2-PIX-FMT-HM12"> <entry><constant>V4L2_PIX_FMT_HM12</constant></entry> <entry>'HM12'</entry> <entry><para>YUV 4:2:0 format used by the IVTV driver, <ulink url="http://www.ivtvdriver.org/"> http://www.ivtvdriver.org/</ulink></para><para>The format is documented in the kernel sources in the file <filename>Documentation/video4linux/cx2341x/README.hm12</filename> </para></entry> </row> <row id="V4L2-PIX-FMT-CPIA1"> <entry><constant>V4L2_PIX_FMT_CPIA1</constant></entry> <entry>'CPIA'</entry> <entry>YUV format used by the gspca cpia1 driver.</entry> </row> <row id="V4L2-PIX-FMT-JPGL"> <entry><constant>V4L2_PIX_FMT_JPGL</constant></entry> <entry>'JPGL'</entry> <entry>JPEG-Light format (Pegasus Lossless JPEG) used in Divio webcams NW 80x.</entry> </row> <row id="V4L2-PIX-FMT-SPCA501"> <entry><constant>V4L2_PIX_FMT_SPCA501</constant></entry> <entry>'S501'</entry> <entry>YUYV per line used by the gspca driver.</entry> </row> <row id="V4L2-PIX-FMT-SPCA505"> <entry><constant>V4L2_PIX_FMT_SPCA505</constant></entry> <entry>'S505'</entry> <entry>YYUV per line used by the gspca driver.</entry> </row> <row id="V4L2-PIX-FMT-SPCA508"> <entry><constant>V4L2_PIX_FMT_SPCA508</constant></entry> <entry>'S508'</entry> <entry>YUVY per line used by the gspca driver.</entry> </row> <row id="V4L2-PIX-FMT-SPCA561"> <entry><constant>V4L2_PIX_FMT_SPCA561</constant></entry> <entry>'S561'</entry> <entry>Compressed GBRG Bayer format used by the gspca driver.</entry> </row> <row id="V4L2-PIX-FMT-PAC207"> <entry><constant>V4L2_PIX_FMT_PAC207</constant></entry> <entry>'P207'</entry> <entry>Compressed BGGR Bayer format used by the gspca driver.</entry> </row> <row id="V4L2-PIX-FMT-MR97310A"> <entry><constant>V4L2_PIX_FMT_MR97310A</constant></entry> <entry>'M310'</entry> <entry>Compressed BGGR Bayer format used by the gspca driver.</entry> </row> <row id="V4L2-PIX-FMT-JL2005BCD"> <entry><constant>V4L2_PIX_FMT_JL2005BCD</constant></entry> <entry>'JL20'</entry> <entry>JPEG compressed RGGB Bayer format used by the gspca driver.</entry> </row> <row id="V4L2-PIX-FMT-OV511"> <entry><constant>V4L2_PIX_FMT_OV511</constant></entry> <entry>'O511'</entry> <entry>OV511 JPEG format used by the gspca driver.</entry> </row> <row id="V4L2-PIX-FMT-OV518"> <entry><constant>V4L2_PIX_FMT_OV518</constant></entry> <entry>'O518'</entry> <entry>OV518 JPEG format used by the gspca driver.</entry> </row> <row id="V4L2-PIX-FMT-PJPG"> <entry><constant>V4L2_PIX_FMT_PJPG</constant></entry> <entry>'PJPG'</entry> <entry>Pixart 73xx JPEG format used by the gspca driver.</entry> </row> <row id="V4L2-PIX-FMT-SE401"> <entry><constant>V4L2_PIX_FMT_SE401</constant></entry> <entry>'S401'</entry> <entry>Compressed RGB format used by the gspca se401 driver</entry> </row> <row id="V4L2-PIX-FMT-SQ905C"> <entry><constant>V4L2_PIX_FMT_SQ905C</constant></entry> <entry>'905C'</entry> <entry>Compressed RGGB bayer format used by the gspca driver.</entry> </row> <row id="V4L2-PIX-FMT-MJPEG"> <entry><constant>V4L2_PIX_FMT_MJPEG</constant></entry> <entry>'MJPG'</entry> <entry>Compressed format used by the Zoran driver</entry> </row> <row id="V4L2-PIX-FMT-PWC1"> <entry><constant>V4L2_PIX_FMT_PWC1</constant></entry> <entry>'PWC1'</entry> <entry>Compressed format of the PWC driver.</entry> </row> <row id="V4L2-PIX-FMT-PWC2"> <entry><constant>V4L2_PIX_FMT_PWC2</constant></entry> <entry>'PWC2'</entry> <entry>Compressed format of the PWC driver.</entry> </row> <row id="V4L2-PIX-FMT-SN9C10X"> <entry><constant>V4L2_PIX_FMT_SN9C10X</constant></entry> <entry>'S910'</entry> <entry>Compressed format of the SN9C102 driver.</entry> </row> <row id="V4L2-PIX-FMT-SN9C20X-I420"> <entry><constant>V4L2_PIX_FMT_SN9C20X_I420</constant></entry> <entry>'S920'</entry> <entry>YUV 4:2:0 format of the gspca sn9c20x driver.</entry> </row> <row id="V4L2-PIX-FMT-SN9C2028"> <entry><constant>V4L2_PIX_FMT_SN9C2028</constant></entry> <entry>'SONX'</entry> <entry>Compressed GBRG bayer format of the gspca sn9c2028 driver.</entry> </row> <row id="V4L2-PIX-FMT-STV0680"> <entry><constant>V4L2_PIX_FMT_STV0680</constant></entry> <entry>'S680'</entry> <entry>Bayer format of the gspca stv0680 driver.</entry> </row> <row id="V4L2-PIX-FMT-WNVA"> <entry><constant>V4L2_PIX_FMT_WNVA</constant></entry> <entry>'WNVA'</entry> <entry><para>Used by the Winnov Videum driver, <ulink url="http://www.thedirks.org/winnov/"> http://www.thedirks.org/winnov/</ulink></para></entry> </row> <row id="V4L2-PIX-FMT-TM6000"> <entry><constant>V4L2_PIX_FMT_TM6000</constant></entry> <entry>'TM60'</entry> <entry><para>Used by Trident tm6000</para></entry> </row> <row id="V4L2-PIX-FMT-CIT-YYVYUY"> <entry><constant>V4L2_PIX_FMT_CIT_YYVYUY</constant></entry> <entry>'CITV'</entry> <entry><para>Used by xirlink CIT, found at IBM webcams.</para> <para>Uses one line of Y then 1 line of VYUY</para> </entry> </row> <row id="V4L2-PIX-FMT-KONICA420"> <entry><constant>V4L2_PIX_FMT_KONICA420</constant></entry> <entry>'KONI'</entry> <entry><para>Used by Konica webcams.</para> <para>YUV420 planar in blocks of 256 pixels.</para> </entry> </row> <row id="V4L2-PIX-FMT-YYUV"> <entry><constant>V4L2_PIX_FMT_YYUV</constant></entry> <entry>'YYUV'</entry> <entry>unknown</entry> </row> <row id="V4L2-PIX-FMT-Y4"> <entry><constant>V4L2_PIX_FMT_Y4</constant></entry> <entry>'Y04 '</entry> <entry>Old 4-bit greyscale format. Only the most significant 4 bits of each byte are used, the other bits are set to 0.</entry> </row> <row id="V4L2-PIX-FMT-Y6"> <entry><constant>V4L2_PIX_FMT_Y6</constant></entry> <entry>'Y06 '</entry> <entry>Old 6-bit greyscale format. Only the most significant 6 bits of each byte are used, the other bits are set to 0.</entry> </row> <row id="V4L2-PIX-FMT-S5C-UYVY-JPG"> <entry><constant>V4L2_PIX_FMT_S5C_UYVY_JPG</constant></entry> <entry>'S5CI'</entry> <entry>Two-planar format used by Samsung S5C73MX cameras. The first plane contains interleaved JPEG and UYVY image data, followed by meta data in form of an array of offsets to the UYVY data blocks. The actual pointer array follows immediately the interleaved JPEG/UYVY data, the number of entries in this array equals the height of the UYVY image. Each entry is a 4-byte unsigned integer in big endian order and it's an offset to a single pixel line of the UYVY image. The first plane can start either with JPEG or UYVY data chunk. The size of a single UYVY block equals the UYVY image's width multiplied by 2. The size of a JPEG chunk depends on the image and can vary with each line. <para>The second plane, at an offset of 4084 bytes, contains a 4-byte offset to the pointer array in the first plane. This offset is followed by a 4-byte value indicating size of the pointer array. All numbers in the second plane are also in big endian order. Remaining data in the second plane is undefined. The information in the second plane allows to easily find location of the pointer array, which can be different for each frame. The size of the pointer array is constant for given UYVY image height.</para> <para>In order to extract UYVY and JPEG frames an application can initially set a data pointer to the start of first plane and then add an offset from the first entry of the pointers table. Such a pointer indicates start of an UYVY image pixel line. Whole UYVY line can be copied to a separate buffer. These steps should be repeated for each line, i.e. the number of entries in the pointer array. Anything what's in between the UYVY lines is JPEG data and should be concatenated to form the JPEG stream. </para> </entry> </row> </tbody> </tgroup> </table> </section>