
This article needs additional citations for verification.(March 2023) |
In digital photography, computer-generated imagery, and colorimetry, a greyscale (more common in Commonwealth English) or grayscale (more common in American English) image is one in which the value of each pixel is a single sample representing only an amount of light; that is, it carries only intensity information. Grayscale images, are black-and-white or gray monochrome, and composed exclusively of shades of gray. The contrast ranges from black at the weakest intensity to white at the strongest.

Grayscale images are distinct from one-bit bi-tonal black-and-white images, which, in the context of computer imaging, are images with only two colors: black and white (also called bilevel or binary images). Grayscale images have many shades of gray in between.
Grayscale images can be the result of measuring the intensity of light at each pixel according to a particular weighted combination of frequencies (or wavelengths), and in such cases they are monochromatic proper when only a single frequency (in practice, a narrow band of frequencies) is captured. The frequencies can in principle be from anywhere in the electromagnetic spectrum (e.g. infrared, visible light, ultraviolet, etc.).
A colorimetric (or more specifically photometric) grayscale image is an image that has a defined grayscale colorspace, which maps the stored numeric sample values to the achromatic channel of a standard colorspace, which itself is based on measured properties of human vision.
If the original color image has no defined colorspace, or if the grayscale image is not intended to have the same human-perceived achromatic intensity as the color image, then there is no unique mapping from such a color image to a grayscale image.
Numerical representations
The intensity of a pixel is expressed within a given range between a minimum and a maximum, inclusive. This range is represented in an abstract way as a range from 0 (or 0%) (total absence, black) and 1 (or 100%) (total presence, white), with any fractional values in between. This notation is used in academic papers, but this does not define what "black" or "white" is in terms of colorimetry. Sometimes the scale is reversed, as in printing where the numeric intensity denotes how much ink is employed in halftoning, with 0% representing the paper white (no ink) and 100% being a solid black (full ink).
In computing, although the grayscale can be computed through rational numbers, image pixels are usually quantized to store them as unsigned integers, to reduce the required storage and computation. Some early grayscale monitors can only display up to sixteen different shades, which would be stored in binary form using 4 bits.[citation needed] But today grayscale images intended for visual display are commonly stored with 8 bits per sampled pixel. This pixel depth allows 256 different intensities (i.e., shades of gray) to be recorded, and also simplifies computation as each pixel sample can be accessed individually as one full byte. However, if these intensities were spaced equally in proportion to the amount of physical light they represent at that pixel (called a linear encoding or scale), the differences between adjacent dark shades could be quite noticeable as banding artifacts, while many of the lighter shades would be "wasted" by encoding a lot of perceptually-indistinguishable increments. Therefore, the shades are instead typically spread out evenly on a gamma-compressed nonlinear scale, which better approximates uniform perceptual increments for both dark and light shades, usually making these 256 shades enough to avoid noticeable increments.
Technical uses (e.g. in medical imaging or remote sensing applications) often require more levels, to make full use of the sensor accuracy (typically 10 or 12 bits per sample) and to reduce rounding errors in computations. Sixteen bits per sample (65,536 levels) is often a convenient choice for such uses, as computers manage 16-bit words efficiently. The TIFF and PNG (among other) image file formats support 16-bit grayscale natively, although browsers and many imaging programs tend to ignore the low order 8 bits of each pixel. Internally for computation and working storage, image processing software typically uses integer or floating-point numbers of size 16 or 32 bits.
Converting color to grayscale
Conversion of an arbitrary color image to grayscale is not unique in general; different weighting of the color channels effectively represent the effect of shooting black-and-white film with different-colored photographic filters on the cameras.
Colorimetric (perceptual luminance-preserving) conversion to grayscale
A common strategy is to use the principles of photometry or, more broadly, colorimetry to calculate the grayscale values (in the target grayscale colorspace) so as to have the same luminance (technically relative luminance) as the original color image (according to its colorspace). In addition to the same (relative) luminance, this method also ensures that both images will have the same absolute luminance when displayed, as can be measured by instruments in its SI units of candelas per square meter, in any given area of the image, given equal whitepoints. Luminance itself is defined using a standard model of human vision, so preserving the luminance in the grayscale image also preserves other perceptual lightness measures, such as L* (as in the 1976 CIE Lab color space) which is determined by the linear luminance Y itself (as in the CIE 1931 XYZ color space) which we will refer to here as Ylinear to avoid any ambiguity.
To convert a color from a colorspace based on a typical gamma-compressed (nonlinear) RGB color model to a grayscale representation of its luminance, the gamma compression function must first be removed via gamma expansion (linearization) to transform the image to a linear RGB colorspace, so that the appropriate weighted sum can be applied to the linear color components () to calculate the linear luminance Ylinear, which can then be gamma-compressed back again if the grayscale result is also to be encoded and stored in a typical nonlinear colorspace.
For the common sRGB color space, gamma expansion is defined as
where Csrgb represents any of the three gamma-compressed sRGB primaries (Rsrgb, Gsrgb, and Bsrgb, each in range [0,1]) and Clinear is the corresponding linear-intensity value (Rlinear, Glinear, and Blinear, also in range [0,1]). Then, linear luminance is calculated as a weighted sum of the three linear-intensity values. The sRGB color space is defined in terms of the CIE 1931 linear luminance Ylinear, which is given by
These three particular coefficients represent the intensity (luminance) perception of typical trichromat humans to light of the precise Rec. 709 additive primary colors (chromaticities) that are used in the definition of sRGB. Human vision is most sensitive to green, so this has the greatest coefficient value (0.7152), and least sensitive to blue, so this has the smallest coefficient (0.0722). To encode grayscale intensity in linear RGB, each of the three color components can be set to equal the calculated linear luminance (replacing
by the values
to get this linear grayscale), which then typically needs to be gamma compressed to get back to a conventional non-linear representation. For sRGB, each of its three primaries is then set to the same gamma-compressed Ysrgb given by the inverse of the gamma expansion above as
Because the three sRGB components are then equal, indicating that it is actually a gray image (not color), it is only necessary to store these values once, and we call this the resulting grayscale image. This is how it will normally be stored in sRGB-compatible image formats that support a single-channel grayscale representation, such as JPEG or PNG. Web browsers and other software that recognizes sRGB images should produce the same rendering for such a grayscale image as it would for a "color" sRGB image having the same values in all three color channels.
Luma coding in video systems
For images in color spaces such as Y'UV and its relatives, which are used in standard color TV and video systems such as PAL, SECAM, and NTSC, a nonlinear luma component (Y′) is calculated directly from gamma-compressed primary intensities as a weighted sum, which, although not a perfect representation of the colorimetric luminance, can be calculated more quickly without the gamma expansion and compression used in photometric/colorimetric calculations. In the Y'UV and Y'IQ models used by PAL and NTSC, the rec601 luma (Y′) component is computed as where we use the prime to distinguish these nonlinear values from the sRGB nonlinear values (discussed above) which use a somewhat different gamma compression formula, and from the linear RGB components. The ITU-R BT.709 standard used for HDTV developed by the ATSC uses different color coefficients, computing the luma component as
Although these are numerically the same coefficients used in sRGB above, the effect is different because here they are being applied directly to gamma-compressed values rather than to the linearized values. The ITU-R BT.2100 standard for HDR television uses yet different coefficients, computing the luma component as
Normally these colorspaces are transformed back to nonlinear R'G'B' before rendering for viewing. To the extent that enough precision remains, they can then be rendered accurately.
But if the luma component Y' itself is instead used directly as a grayscale representation of the color image, luminance is not preserved: two colors can have the same luma Y′ but different CIE linear luminance Y (and thus different nonlinear Ysrgb as defined above) and therefore appear darker or lighter to a typical human than the original color. Similarly, two colors having the same luminance Y (and thus the same Ysrgb) will in general have different luma by either of the Y′ luma definitions above.
Grayscale as single channels of multichannel color images
This section does not cite any sources.(March 2023) |
Color images are often built of several stacked color channels, each of them representing value levels of the given channel. For example, RGB images are composed of three independent channels for red, green and blue primary color components; CMYK images have four channels for cyan, magenta, yellow and black ink plates, etc.
Here is an example of color channel splitting of a full RGB color image. The column at left shows the isolated color channels in natural colors, while at right there are their grayscale equivalences:
The reverse is also possible: to build a full-color image from their separate grayscale channels. By mangling channels, using offsets, rotating and other manipulations, artistic effects can be achieved instead of accurately reproducing the original image.
See also
- Channel (digital image)
- Halftone
- Duotone
- False-color
- Sepia tone
- Cyanotype
- Morphological image processing
- Mezzotint
- List of monochrome and RGB color formats – Monochrome palettes section
- List of software palettes – Color gradient palettes and false color palettes sections
- Achromatopsia, total color blindness, in which vision is limited to a grayscale
- Zone System
References
- Johnson, Stephen (2006). Stephen Johnson on Digital Photography. O'Reilly. ISBN 0-596-52370-X.
- Poynton, Charles (2012). Digital Video and HD: Algorithms and Interfaces (2nd ed.). Morgan Kaufmann. pp. 31–35, 65–68, 333, 337. ISBN 978-0-12-391926-7. Retrieved 2022-03-31.
- Poynton, Charles A. (2022-03-14). Written at San Jose, Calif.. Rogowitz, B. E.; Pappas, T. N. (eds.). Rehabilitation of Gamma (PDF). SPIE/IS&T Conference 3299: Human Vision and Electronic Imaging III; January 26–30, 1998. Bellingham, Wash.: SPIE. doi:10.1117/12.320126. Archived (PDF) from the original on 2023-04-23.
- Poynton, Charles A. (2004-02-25). "Constant Luminance". Video Engineering. Archived from the original on 2023-03-16.
- Lindbloom, Bruce (2017-04-06). "RGB Working Space Information". Archived from the original on 2023-06-01.
- Stokes, Michael; Anderson, Matthew; Chandrasekar, Srinivasan; Motta, Ricardo (1996-11-05). "A Standard Default Color Space for the Internet – sRGB". World Wide Web Consortium – Graphics on the Web. Part 2, matrix in equation 1.8. Archived from the original on 2023-05-24.
- Burger, Wilhelm; Burge, Mark J. (2010). Principles of Digital Image Processing Core Algorithms. Springer Science & Business Media. pp. 110–111. ISBN 978-1-84800-195-4.
- Poynton, Charles A. (1997-07-15). "The Magnitude of Nonconstant Luminance Errors" (PDF).
This article needs additional citations for verification Please help improve this article by adding citations to reliable sources Unsourced material may be challenged and removed Find sources Grayscale news newspapers books scholar JSTOR March 2023 Learn how and when to remove this message In digital photography computer generated imagery and colorimetry a greyscale more common in Commonwealth English or grayscale more common in American English image is one in which the value of each pixel is a single sample representing only an amount of light that is it carries only intensity information Grayscale images are black and white or gray monochrome and composed exclusively of shades of gray The contrast ranges from black at the weakest intensity to white at the strongest Grayscale image of a parrot Grayscale images are distinct from one bit bi tonal black and white images which in the context of computer imaging are images with only two colors black and white also called bilevel or binary images Grayscale images have many shades of gray in between Grayscale images can be the result of measuring the intensity of light at each pixel according to a particular weighted combination of frequencies or wavelengths and in such cases they are monochromatic proper when only a single frequency in practice a narrow band of frequencies is captured The frequencies can in principle be from anywhere in the electromagnetic spectrum e g infrared visible light ultraviolet etc A colorimetric or more specifically photometric grayscale image is an image that has a defined grayscale colorspace which maps the stored numeric sample values to the achromatic channel of a standard colorspace which itself is based on measured properties of human vision If the original color image has no defined colorspace or if the grayscale image is not intended to have the same human perceived achromatic intensity as the color image then there is no unique mapping from such a color image to a grayscale image Numerical representations The intensity of a pixel is expressed within a given range between a minimum and a maximum inclusive This range is represented in an abstract way as a range from 0 or 0 total absence black and 1 or 100 total presence white with any fractional values in between This notation is used in academic papers but this does not define what black or white is in terms of colorimetry Sometimes the scale is reversed as in printing where the numeric intensity denotes how much ink is employed in halftoning with 0 representing the paper white no ink and 100 being a solid black full ink In computing although the grayscale can be computed through rational numbers image pixels are usually quantized to store them as unsigned integers to reduce the required storage and computation Some early grayscale monitors can only display up to sixteen different shades which would be stored in binary form using 4 bits citation needed But today grayscale images intended for visual display are commonly stored with 8 bits per sampled pixel This pixel depth allows 256 different intensities i e shades of gray to be recorded and also simplifies computation as each pixel sample can be accessed individually as one full byte However if these intensities were spaced equally in proportion to the amount of physical light they represent at that pixel called a linear encoding or scale the differences between adjacent dark shades could be quite noticeable as banding artifacts while many of the lighter shades would be wasted by encoding a lot of perceptually indistinguishable increments Therefore the shades are instead typically spread out evenly on a gamma compressed nonlinear scale which better approximates uniform perceptual increments for both dark and light shades usually making these 256 shades enough to avoid noticeable increments Technical uses e g in medical imaging or remote sensing applications often require more levels to make full use of the sensor accuracy typically 10 or 12 bits per sample and to reduce rounding errors in computations Sixteen bits per sample 65 536 levels is often a convenient choice for such uses as computers manage 16 bit words efficiently The TIFF and PNG among other image file formats support 16 bit grayscale natively although browsers and many imaging programs tend to ignore the low order 8 bits of each pixel Internally for computation and working storage image processing software typically uses integer or floating point numbers of size 16 or 32 bits Converting color to grayscaleExamples of conversion from a full color image to grayscale using Adobe Photoshop s Channel Mixer compared to the original image and colorimetric conversion to grayscale Conversion of an arbitrary color image to grayscale is not unique in general different weighting of the color channels effectively represent the effect of shooting black and white film with different colored photographic filters on the cameras Colorimetric perceptual luminance preserving conversion to grayscale A common strategy is to use the principles of photometry or more broadly colorimetry to calculate the grayscale values in the target grayscale colorspace so as to have the same luminance technically relative luminance as the original color image according to its colorspace In addition to the same relative luminance this method also ensures that both images will have the same absolute luminance when displayed as can be measured by instruments in its SI units of candelas per square meter in any given area of the image given equal whitepoints Luminance itself is defined using a standard model of human vision so preserving the luminance in the grayscale image also preserves other perceptual lightness measures such as L as in the 1976 CIE Lab color space which is determined by the linear luminance Y itself as in the CIE 1931 XYZ color space which we will refer to here as Ylinear to avoid any ambiguity To convert a color from a colorspace based on a typical gamma compressed nonlinear RGB color model to a grayscale representation of its luminance the gamma compression function must first be removed via gamma expansion linearization to transform the image to a linear RGB colorspace so that the appropriate weighted sum can be applied to the linear color components Rlinear Glinear Blinear displaystyle R mathrm linear G mathrm linear B mathrm linear to calculate the linear luminance Ylinear which can then be gamma compressed back again if the grayscale result is also to be encoded and stored in a typical nonlinear colorspace For the common sRGB color space gamma expansion is defined as Clinear Csrgb12 92 if Csrgb 0 04045 Csrgb 0 0551 055 2 4 otherwise displaystyle C mathrm linear begin cases frac C mathrm srgb 12 92 amp text if C mathrm srgb leq 0 04045 left frac C mathrm srgb 0 055 1 055 right 2 4 amp text otherwise end cases where Csrgb represents any of the three gamma compressed sRGB primaries Rsrgb Gsrgb and Bsrgb each in range 0 1 and Clinear is the corresponding linear intensity value Rlinear Glinear and Blinear also in range 0 1 Then linear luminance is calculated as a weighted sum of the three linear intensity values The sRGB color space is defined in terms of the CIE 1931 linear luminance Ylinear which is given by Ylinear 0 2126Rlinear 0 7152Glinear 0 0722Blinear displaystyle Y mathrm linear 0 2126R mathrm linear 0 7152G mathrm linear 0 0722B mathrm linear These three particular coefficients represent the intensity luminance perception of typical trichromat humans to light of the precise Rec 709 additive primary colors chromaticities that are used in the definition of sRGB Human vision is most sensitive to green so this has the greatest coefficient value 0 7152 and least sensitive to blue so this has the smallest coefficient 0 0722 To encode grayscale intensity in linear RGB each of the three color components can be set to equal the calculated linear luminance Ylinear displaystyle Y mathrm linear replacing Rlinear Glinear Blinear displaystyle R mathrm linear G mathrm linear B mathrm linear by the values Ylinear Ylinear Ylinear displaystyle Y mathrm linear Y mathrm linear Y mathrm linear to get this linear grayscale which then typically needs to be gamma compressed to get back to a conventional non linear representation For sRGB each of its three primaries is then set to the same gamma compressed Ysrgb given by the inverse of the gamma expansion above as Ysrgb 12 92 Ylinear if Ylinear 0 00313081 055 Ylinear1 2 4 0 055 otherwise displaystyle Y mathrm srgb begin cases 12 92 Y mathrm linear amp text if Y mathrm linear leq 0 0031308 1 055 Y mathrm linear 1 2 4 0 055 amp text otherwise end cases Because the three sRGB components are then equal indicating that it is actually a gray image not color it is only necessary to store these values once and we call this the resulting grayscale image This is how it will normally be stored in sRGB compatible image formats that support a single channel grayscale representation such as JPEG or PNG Web browsers and other software that recognizes sRGB images should produce the same rendering for such a grayscale image as it would for a color sRGB image having the same values in all three color channels Luma coding in video systems For images in color spaces such as Y UV and its relatives which are used in standard color TV and video systems such as PAL SECAM and NTSC a nonlinear luma component Y is calculated directly from gamma compressed primary intensities as a weighted sum which although not a perfect representation of the colorimetric luminance can be calculated more quickly without the gamma expansion and compression used in photometric colorimetric calculations In the Y UV and Y IQ models used by PAL and NTSC the rec601 luma Y component is computed as Y 0 299R 0 587G 0 114B displaystyle Y 0 299R 0 587G 0 114B where we use the prime to distinguish these nonlinear values from the sRGB nonlinear values discussed above which use a somewhat different gamma compression formula and from the linear RGB components The ITU R BT 709 standard used for HDTV developed by the ATSC uses different color coefficients computing the luma component as Y 0 2126R 0 7152G 0 0722B displaystyle Y 0 2126R 0 7152G 0 0722B Although these are numerically the same coefficients used in sRGB above the effect is different because here they are being applied directly to gamma compressed values rather than to the linearized values The ITU R BT 2100 standard for HDR television uses yet different coefficients computing the luma component as Y 0 2627R 0 6780G 0 0593B displaystyle Y 0 2627R 0 6780G 0 0593B Normally these colorspaces are transformed back to nonlinear R G B before rendering for viewing To the extent that enough precision remains they can then be rendered accurately But if the luma component Y itself is instead used directly as a grayscale representation of the color image luminance is not preserved two colors can have the same luma Y but different CIE linear luminance Y and thus different nonlinear Ysrgb as defined above and therefore appear darker or lighter to a typical human than the original color Similarly two colors having the same luminance Y and thus the same Ysrgb will in general have different luma by either of the Y luma definitions above Grayscale as single channels of multichannel color imagesThis section does not cite any sources Please help improve this section by adding citations to reliable sources Unsourced material may be challenged and removed March 2023 Learn how and when to remove this message Color images are often built of several stacked color channels each of them representing value levels of the given channel For example RGB images are composed of three independent channels for red green and blue primary color components CMYK images have four channels for cyan magenta yellow and black ink plates etc Here is an example of color channel splitting of a full RGB color image The column at left shows the isolated color channels in natural colors while at right there are their grayscale equivalences Composition of RGB from three grayscale images The reverse is also possible to build a full color image from their separate grayscale channels By mangling channels using offsets rotating and other manipulations artistic effects can be achieved instead of accurately reproducing the original image See alsoChannel digital image Halftone Duotone False color Sepia tone Cyanotype Morphological image processing Mezzotint List of monochrome and RGB color formats Monochrome palettes section List of software palettes Color gradient palettes and false color palettes sections Achromatopsia total color blindness in which vision is limited to a grayscale Zone SystemReferencesJohnson Stephen 2006 Stephen Johnson on Digital Photography O Reilly ISBN 0 596 52370 X Poynton Charles 2012 Digital Video and HD Algorithms and Interfaces 2nd ed Morgan Kaufmann pp 31 35 65 68 333 337 ISBN 978 0 12 391926 7 Retrieved 2022 03 31 Poynton Charles A 2022 03 14 Written at San Jose Calif Rogowitz B E Pappas T N eds Rehabilitation of Gamma PDF SPIE IS amp T Conference 3299 Human Vision and Electronic Imaging III January 26 30 1998 Bellingham Wash SPIE doi 10 1117 12 320126 Archived PDF from the original on 2023 04 23 Poynton Charles A 2004 02 25 Constant Luminance Video Engineering Archived from the original on 2023 03 16 Lindbloom Bruce 2017 04 06 RGB Working Space Information Archived from the original on 2023 06 01 Stokes Michael Anderson Matthew Chandrasekar Srinivasan Motta Ricardo 1996 11 05 A Standard Default Color Space for the Internet sRGB World Wide Web Consortium Graphics on the Web Part 2 matrix in equation 1 8 Archived from the original on 2023 05 24 Burger Wilhelm Burge Mark J 2010 Principles of Digital Image Processing Core Algorithms Springer Science amp Business Media pp 110 111 ISBN 978 1 84800 195 4 Poynton Charles A 1997 07 15 The Magnitude of Nonconstant Luminance Errors PDF