All Downloads are FREE. Search and download functionalities are using the official Maven repository.

android.graphics.ImageFormat Maven / Gradle / Ivy

There is a newer version: 0.3.3
Show newest version
/*
 * Copyright (C) 2010 The Android Open Source Project
 *
 * Licensed under the Apache License, Version 2.0 (the "License");
 * you may not use this file except in compliance with the License.
 * You may obtain a copy of the License at
 *
 *      http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */

package android.graphics;

public class ImageFormat {
    /*
     * these constants are chosen to be binary compatible with their previous
     * location in PixelFormat.java
     */

    public static final int UNKNOWN = 0;

    /**
     * RGB format used for pictures encoded as RGB_565. See
     * {@link android.hardware.Camera.Parameters#setPictureFormat(int)}.
     */
    public static final int RGB_565 = 4;

    /**
     * 

Android YUV format.

* *

This format is exposed to software decoders and applications.

* *

YV12 is a 4:2:0 YCrCb planar format comprised of a WxH Y plane followed * by (W/2) x (H/2) Cr and Cb planes.

* *

This format assumes *

    *
  • an even width
  • *
  • an even height
  • *
  • a horizontal stride multiple of 16 pixels
  • *
  • a vertical stride equal to the height
  • *
*

* *
 y_size = stride * height
     * c_stride = ALIGN(stride/2, 16)
     * c_size = c_stride * height/2
     * size = y_size + c_size * 2
     * cr_offset = y_size
     * cb_offset = y_size + c_size
* *

For the {@link android.hardware.camera2} API, the {@link #YUV_420_888} format is * recommended for YUV output instead.

* *

For the older camera API, this format is guaranteed to be supported for * {@link android.hardware.Camera} preview images since API level 12; for earlier API versions, * check {@link android.hardware.Camera.Parameters#getSupportedPreviewFormats()}. * *

Note that for camera preview callback use (see * {@link android.hardware.Camera#setPreviewCallback}), the * stride value is the smallest possible; that is, it is equal * to: * *

stride = ALIGN(width, 16)
* * @see android.hardware.Camera.Parameters#setPreviewCallback * @see android.hardware.Camera.Parameters#setPreviewFormat * @see android.hardware.Camera.Parameters#getSupportedPreviewFormats *

*/ public static final int YV12 = 0x32315659; /** *

Android Y8 format.

* *

Y8 is a YUV planar format comprised of a WxH Y plane only, with each pixel * being represented by 8 bits. It is equivalent to just the Y plane from {@link #YV12} * format.

* *

This format assumes *

    *
  • an even width
  • *
  • an even height
  • *
  • a horizontal stride multiple of 16 pixels
  • *
*

* *
 y_size = stride * height 
* *

For example, the {@link android.media.Image} object can provide data * in this format from a {@link android.hardware.camera2.CameraDevice} * through a {@link android.media.ImageReader} object if this format is * supported by {@link android.hardware.camera2.CameraDevice}.

* * @see android.media.Image * @see android.media.ImageReader * @see android.hardware.camera2.CameraDevice * * @hide */ public static final int Y8 = 0x20203859; /** *

Android Y16 format.

* * Y16 is a YUV planar format comprised of a WxH Y plane, with each pixel * being represented by 16 bits. It is just like {@link #Y8}, but has 16 * bits per pixel (little endian).

* *

This format assumes *

    *
  • an even width
  • *
  • an even height
  • *
  • a horizontal stride multiple of 16 pixels
  • *
*

* *
 y_size = stride * height 
* *

For example, the {@link android.media.Image} object can provide data * in this format from a {@link android.hardware.camera2.CameraDevice} * through a {@link android.media.ImageReader} object if this format is * supported by {@link android.hardware.camera2.CameraDevice}.

* * @see android.media.Image * @see android.media.ImageReader * @see android.hardware.camera2.CameraDevice * * @hide */ public static final int Y16 = 0x20363159; /** * YCbCr format, used for video. * *

For the {@link android.hardware.camera2} API, the {@link #YUV_420_888} format is * recommended for YUV output instead.

* *

Whether this format is supported by the old camera API can be determined by * {@link android.hardware.Camera.Parameters#getSupportedPreviewFormats()}.

* */ public static final int NV16 = 0x10; /** * YCrCb format used for images, which uses the NV21 encoding format. * *

This is the default format * for {@link android.hardware.Camera} preview images, when not otherwise set with * {@link android.hardware.Camera.Parameters#setPreviewFormat(int)}.

* *

For the {@link android.hardware.camera2} API, the {@link #YUV_420_888} format is * recommended for YUV output instead.

*/ public static final int NV21 = 0x11; /** * YCbCr format used for images, which uses YUYV (YUY2) encoding format. * *

For the {@link android.hardware.camera2} API, the {@link #YUV_420_888} format is * recommended for YUV output instead.

* *

This is an alternative format for {@link android.hardware.Camera} preview images. Whether * this format is supported by the camera hardware can be determined by * {@link android.hardware.Camera.Parameters#getSupportedPreviewFormats()}.

*/ public static final int YUY2 = 0x14; /** * Compressed JPEG format. * *

This format is always supported as an output format for the * {@link android.hardware.camera2} API, and as a picture format for the older * {@link android.hardware.Camera} API

*/ public static final int JPEG = 0x100; /** *

Multi-plane Android YUV 420 format

* *

This format is a generic YCbCr format, capable of describing any 4:2:0 * chroma-subsampled planar or semiplanar buffer (but not fully interleaved), * with 8 bits per color sample.

* *

Images in this format are always represented by three separate buffers * of data, one for each color plane. Additional information always * accompanies the buffers, describing the row stride and the pixel stride * for each plane.

* *

The order of planes in the array returned by * {@link android.media.Image#getPlanes() Image#getPlanes()} is guaranteed such that * plane #0 is always Y, plane #1 is always U (Cb), and plane #2 is always V (Cr).

* *

The Y-plane is guaranteed not to be interleaved with the U/V planes * (in particular, pixel stride is always 1 in * {@link android.media.Image.Plane#getPixelStride() yPlane.getPixelStride()}).

* *

The U/V planes are guaranteed to have the same row stride and pixel stride * (in particular, * {@link android.media.Image.Plane#getRowStride() uPlane.getRowStride()} * == {@link android.media.Image.Plane#getRowStride() vPlane.getRowStride()} and * {@link android.media.Image.Plane#getPixelStride() uPlane.getPixelStride()} * == {@link android.media.Image.Plane#getPixelStride() vPlane.getPixelStride()}; * ).

* *

For example, the {@link android.media.Image} object can provide data * in this format from a {@link android.hardware.camera2.CameraDevice} * through a {@link android.media.ImageReader} object.

* * @see android.media.Image * @see android.media.ImageReader * @see android.hardware.camera2.CameraDevice */ public static final int YUV_420_888 = 0x23; /** *

Multi-plane Android YUV 422 format

* *

This format is a generic YCbCr format, capable of describing any 4:2:2 * chroma-subsampled (planar, semiplanar or interleaved) format, * with 8 bits per color sample.

* *

Images in this format are always represented by three separate buffers * of data, one for each color plane. Additional information always * accompanies the buffers, describing the row stride and the pixel stride * for each plane.

* *

The order of planes in the array returned by * {@link android.media.Image#getPlanes() Image#getPlanes()} is guaranteed such that * plane #0 is always Y, plane #1 is always U (Cb), and plane #2 is always V (Cr).

* *

In contrast to the {@link #YUV_420_888} format, the Y-plane may have a pixel * stride greater than 1 in * {@link android.media.Image.Plane#getPixelStride() yPlane.getPixelStride()}.

* *

The U/V planes are guaranteed to have the same row stride and pixel stride * (in particular, * {@link android.media.Image.Plane#getRowStride() uPlane.getRowStride()} * == {@link android.media.Image.Plane#getRowStride() vPlane.getRowStride()} and * {@link android.media.Image.Plane#getPixelStride() uPlane.getPixelStride()} * == {@link android.media.Image.Plane#getPixelStride() vPlane.getPixelStride()}; * ).

* *

For example, the {@link android.media.Image} object can provide data * in this format from a {@link android.media.MediaCodec} * through {@link android.media.MediaCodec#getOutputImage} object.

* * @see android.media.Image * @see android.media.MediaCodec */ public static final int YUV_422_888 = 0x27; /** *

Multi-plane Android YUV 444 format

* *

This format is a generic YCbCr format, capable of describing any 4:4:4 * (planar, semiplanar or interleaved) format, * with 8 bits per color sample.

* *

Images in this format are always represented by three separate buffers * of data, one for each color plane. Additional information always * accompanies the buffers, describing the row stride and the pixel stride * for each plane.

* *

The order of planes in the array returned by * {@link android.media.Image#getPlanes() Image#getPlanes()} is guaranteed such that * plane #0 is always Y, plane #1 is always U (Cb), and plane #2 is always V (Cr).

* *

In contrast to the {@link #YUV_420_888} format, the Y-plane may have a pixel * stride greater than 1 in * {@link android.media.Image.Plane#getPixelStride() yPlane.getPixelStride()}.

* *

The U/V planes are guaranteed to have the same row stride and pixel stride * (in particular, * {@link android.media.Image.Plane#getRowStride() uPlane.getRowStride()} * == {@link android.media.Image.Plane#getRowStride() vPlane.getRowStride()} and * {@link android.media.Image.Plane#getPixelStride() uPlane.getPixelStride()} * == {@link android.media.Image.Plane#getPixelStride() vPlane.getPixelStride()}; * ).

* *

For example, the {@link android.media.Image} object can provide data * in this format from a {@link android.media.MediaCodec} * through {@link android.media.MediaCodec#getOutputImage} object.

* * @see android.media.Image * @see android.media.MediaCodec */ public static final int YUV_444_888 = 0x28; /** *

Multi-plane Android RGB format

* *

This format is a generic RGB format, capable of describing most RGB formats, * with 8 bits per color sample.

* *

Images in this format are always represented by three separate buffers * of data, one for each color plane. Additional information always * accompanies the buffers, describing the row stride and the pixel stride * for each plane.

* *

The order of planes in the array returned by * {@link android.media.Image#getPlanes() Image#getPlanes()} is guaranteed such that * plane #0 is always R (red), plane #1 is always G (green), and plane #2 is always B * (blue).

* *

All three planes are guaranteed to have the same row strides and pixel strides.

* *

For example, the {@link android.media.Image} object can provide data * in this format from a {@link android.media.MediaCodec} * through {@link android.media.MediaCodec#getOutputImage} object.

* * @see android.media.Image * @see android.media.MediaCodec */ public static final int FLEX_RGB_888 = 0x29; /** *

Multi-plane Android RGBA format

* *

This format is a generic RGBA format, capable of describing most RGBA formats, * with 8 bits per color sample.

* *

Images in this format are always represented by four separate buffers * of data, one for each color plane. Additional information always * accompanies the buffers, describing the row stride and the pixel stride * for each plane.

* *

The order of planes in the array returned by * {@link android.media.Image#getPlanes() Image#getPlanes()} is guaranteed such that * plane #0 is always R (red), plane #1 is always G (green), plane #2 is always B (blue), * and plane #3 is always A (alpha). This format may represent pre-multiplied or * non-premultiplied alpha.

* *

All four planes are guaranteed to have the same row strides and pixel strides.

* *

For example, the {@link android.media.Image} object can provide data * in this format from a {@link android.media.MediaCodec} * through {@link android.media.MediaCodec#getOutputImage} object.

* * @see android.media.Image * @see android.media.MediaCodec */ public static final int FLEX_RGBA_8888 = 0x2A; /** *

General raw camera sensor image format, usually representing a * single-channel Bayer-mosaic image. Each pixel color sample is stored with * 16 bits of precision.

* *

The layout of the color mosaic, the maximum and minimum encoding * values of the raw pixel data, the color space of the image, and all other * needed information to interpret a raw sensor image must be queried from * the {@link android.hardware.camera2.CameraDevice} which produced the * image.

*/ public static final int RAW_SENSOR = 0x20; /** *

Private raw camera sensor image format, a single channel image with * implementation depedent pixel layout.

* *

RAW_PRIVATE is a format for unprocessed raw image buffers coming from an * image sensor. The actual structure of buffers of this format is * implementation-dependent.

* */ public static final int RAW_PRIVATE = 0x24; /** *

* Android 10-bit raw format *

*

* This is a single-plane, 10-bit per pixel, densely packed (in each row), * unprocessed format, usually representing raw Bayer-pattern images coming * from an image sensor. *

*

* In an image buffer with this format, starting from the first pixel of * each row, each 4 consecutive pixels are packed into 5 bytes (40 bits). * Each one of the first 4 bytes contains the top 8 bits of each pixel, The * fifth byte contains the 2 least significant bits of the 4 pixels, the * exact layout data for each 4 consecutive pixels is illustrated below * ({@code Pi[j]} stands for the jth bit of the ith pixel): *

* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *
bit 7bit 6bit 5bit 4bit 3bit 2bit 1bit 0
Byte 0:P0[9]P0[8]P0[7]P0[6]P0[5]P0[4]P0[3]P0[2]
Byte 1:P1[9]P1[8]P1[7]P1[6]P1[5]P1[4]P1[3]P1[2]
Byte 2:P2[9]P2[8]P2[7]P2[6]P2[5]P2[4]P2[3]P2[2]
Byte 3:P3[9]P3[8]P3[7]P3[6]P3[5]P3[4]P3[3]P3[2]
Byte 4:P3[1]P3[0]P2[1]P2[0]P1[1]P1[0]P0[1]P0[0]
*

* This format assumes *

    *
  • a width multiple of 4 pixels
  • *
  • an even height
  • *
*

* *
size = row stride * height
where the row stride is in bytes, * not pixels. * *

* Since this is a densely packed format, the pixel stride is always 0. The * application must use the pixel data layout defined in above table to * access each row data. When row stride is equal to {@code width * (10 / 8)}, there * will be no padding bytes at the end of each row, the entire image data is * densely packed. When stride is larger than {@code width * (10 / 8)}, padding * bytes will be present at the end of each row. *

*

* For example, the {@link android.media.Image} object can provide data in * this format from a {@link android.hardware.camera2.CameraDevice} (if * supported) through a {@link android.media.ImageReader} object. The * {@link android.media.Image#getPlanes() Image#getPlanes()} will return a * single plane containing the pixel data. The pixel stride is always 0 in * {@link android.media.Image.Plane#getPixelStride()}, and the * {@link android.media.Image.Plane#getRowStride()} describes the vertical * neighboring pixel distance (in bytes) between adjacent rows. *

* * @see android.media.Image * @see android.media.ImageReader * @see android.hardware.camera2.CameraDevice */ public static final int RAW10 = 0x25; /** *

* Android 12-bit raw format *

*

* This is a single-plane, 12-bit per pixel, densely packed (in each row), * unprocessed format, usually representing raw Bayer-pattern images coming * from an image sensor. *

*

* In an image buffer with this format, starting from the first pixel of each * row, each two consecutive pixels are packed into 3 bytes (24 bits). The first * and second byte contains the top 8 bits of first and second pixel. The third * byte contains the 4 least significant bits of the two pixels, the exact layout * data for each two consecutive pixels is illustrated below (Pi[j] stands for * the jth bit of the ith pixel): *

* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *
bit 7bit 6bit 5bit 4bit 3bit 2bit 1bit 0
Byte 0:P0[11]P0[10]P0[ 9]P0[ 8]P0[ 7]P0[ 6]P0[ 5]P0[ 4]
Byte 1:P1[11]P1[10]P1[ 9]P1[ 8]P1[ 7]P1[ 6]P1[ 5]P1[ 4]
Byte 2:P1[ 3]P1[ 2]P1[ 1]P1[ 0]P0[ 3]P0[ 2]P0[ 1]P0[ 0]
*

* This format assumes *

    *
  • a width multiple of 4 pixels
  • *
  • an even height
  • *
*

* *
size = row stride * height
where the row stride is in bytes, * not pixels. * *

* Since this is a densely packed format, the pixel stride is always 0. The * application must use the pixel data layout defined in above table to * access each row data. When row stride is equal to {@code width * (12 / 8)}, there * will be no padding bytes at the end of each row, the entire image data is * densely packed. When stride is larger than {@code width * (12 / 8)}, padding * bytes will be present at the end of each row. *

*

* For example, the {@link android.media.Image} object can provide data in * this format from a {@link android.hardware.camera2.CameraDevice} (if * supported) through a {@link android.media.ImageReader} object. The * {@link android.media.Image#getPlanes() Image#getPlanes()} will return a * single plane containing the pixel data. The pixel stride is always 0 in * {@link android.media.Image.Plane#getPixelStride()}, and the * {@link android.media.Image.Plane#getRowStride()} describes the vertical * neighboring pixel distance (in bytes) between adjacent rows. *

* * @see android.media.Image * @see android.media.ImageReader * @see android.hardware.camera2.CameraDevice */ public static final int RAW12 = 0x26; /** *

Android dense depth image format.

* *

Each pixel is 16 bits, representing a depth ranging measurement from a depth camera or * similar sensor. The 16-bit sample consists of a confidence value and the actual ranging * measurement.

* *

The confidence value is an estimate of correctness for this sample. It is encoded in the * 3 most significant bits of the sample, with a value of 0 representing 100% confidence, a * value of 1 representing 0% confidence, a value of 2 representing 1/7, a value of 3 * representing 2/7, and so on.

* *

As an example, the following sample extracts the range and confidence from the first pixel * of a DEPTH16-format {@link android.media.Image}, and converts the confidence to a * floating-point value between 0 and 1.f inclusive, with 1.f representing maximum confidence: * *

     *    ShortBuffer shortDepthBuffer = img.getPlanes()[0].getBuffer().asShortBuffer();
     *    short depthSample = shortDepthBuffer.get()
     *    short depthRange = (short) (depthSample & 0x1FFF);
     *    short depthConfidence = (short) ((depthSample >> 13) & 0x7);
     *    float depthPercentage = depthConfidence == 0 ? 1.f : (depthConfidence - 1) / 7.f;
     * 
*

* *

This format assumes *

    *
  • an even width
  • *
  • an even height
  • *
  • a horizontal stride multiple of 16 pixels
  • *
*

* *
 y_size = stride * height 
* * When produced by a camera, the units for the range are millimeters. */ public static final int DEPTH16 = 0x44363159; /** * Android sparse depth point cloud format. * *

A variable-length list of 3D points plus a confidence value, with each point represented * by four floats; first the X, Y, Z position coordinates, and then the confidence value.

* *

The number of points is {@code (size of the buffer in bytes) / 16}. * *

The coordinate system and units of the position values depend on the source of the point * cloud data. The confidence value is between 0.f and 1.f, inclusive, with 0 representing 0% * confidence and 1.f representing 100% confidence in the measured position values.

* *

As an example, the following code extracts the first depth point in a DEPTH_POINT_CLOUD * format {@link android.media.Image}: *

     *    FloatBuffer floatDepthBuffer = img.getPlanes()[0].getBuffer().asFloatBuffer();
     *    float x = floatDepthBuffer.get();
     *    float y = floatDepthBuffer.get();
     *    float z = floatDepthBuffer.get();
     *    float confidence = floatDepthBuffer.get();
     * 
* * For camera devices that support the * {@link android.hardware.camera2.CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES_DEPTH_OUTPUT DEPTH_OUTPUT} * capability, DEPTH_POINT_CLOUD coordinates have units of meters, and the coordinate system is * defined by the camera's pose transforms: * {@link android.hardware.camera2.CameraCharacteristics#LENS_POSE_TRANSLATION} and * {@link android.hardware.camera2.CameraCharacteristics#LENS_POSE_ROTATION}. That means the origin is * the optical center of the camera device, and the positive Z axis points along the camera's optical axis, * toward the scene. */ public static final int DEPTH_POINT_CLOUD = 0x101; /** * Unprocessed implementation-dependent raw * depth measurements, opaque with 16 bit * samples. * * @hide */ public static final int RAW_DEPTH = 0x1002; /** * Android private opaque image format. *

* The choices of the actual format and pixel data layout are entirely up to * the device-specific and framework internal implementations, and may vary * depending on use cases even for the same device. The buffers of this * format can be produced by components like * {@link android.media.ImageWriter ImageWriter} , and interpreted correctly * by consumers like {@link android.hardware.camera2.CameraDevice * CameraDevice} based on the device/framework private information. However, * these buffers are not directly accessible to the application. *

*

* When an {@link android.media.Image Image} of this format is obtained from * an {@link android.media.ImageReader ImageReader} or * {@link android.media.ImageWriter ImageWriter}, the * {@link android.media.Image#getPlanes() getPlanes()} method will return an * empty {@link android.media.Image.Plane Plane} array. *

*

* If a buffer of this format is to be used as an OpenGL ES texture, the * framework will assume that sampling the texture will always return an * alpha value of 1.0 (i.e. the buffer contains only opaque pixel values). *

*/ public static final int PRIVATE = 0x22; /** * Use this function to retrieve the number of bits per pixel of an * ImageFormat. * * @param format * @return the number of bits per pixel of the given format or -1 if the * format doesn't exist or is not supported. */ public static int getBitsPerPixel(int format) { switch (format) { case RGB_565: return 16; case NV16: return 16; case YUY2: return 16; case YV12: return 12; case Y8: return 8; case Y16: case DEPTH16: return 16; case NV21: return 12; case YUV_420_888: return 12; case YUV_422_888: return 16; case YUV_444_888: return 24; case FLEX_RGB_888: return 24; case FLEX_RGBA_8888: return 32; case RAW_DEPTH: case RAW_SENSOR: return 16; case RAW10: return 10; case RAW12: return 12; } return -1; } /** * Determine whether or not this is a public-visible {@code format}. * *

In particular, {@code @hide} formats will return {@code false}.

* *

Any other formats (including UNKNOWN) will return {@code false}.

* * @param format an integer format * @return a boolean * * @hide */ public static boolean isPublicFormat(int format) { switch (format) { case RGB_565: case NV16: case YUY2: case YV12: case JPEG: case NV21: case YUV_420_888: case YUV_422_888: case YUV_444_888: case FLEX_RGB_888: case FLEX_RGBA_8888: case RAW_SENSOR: case RAW_PRIVATE: case RAW10: case RAW12: case DEPTH16: case DEPTH_POINT_CLOUD: case PRIVATE: case RAW_DEPTH: return true; } return false; } }




© 2015 - 2024 Weber Informatics LLC | Privacy Policy