![Integer (computer science)](https://www.english.nina.az/image-resize/1600/900/web/wikipedia.jpg)
In computer science, an integer is a datum of integral data type, a data type that represents some range of mathematical integers. Integral data types may be of different sizes and may or may not be allowed to contain negative values. Integers are commonly represented in a computer as a group of binary digits (bits). The size of the grouping varies so the set of integer sizes available varies between different types of computers. Computer hardware nearly always provides a way to represent a processor register or memory address as an integer.
Value and representation
The value of an item with an integral type is the mathematical integer that it corresponds to. Integral types may be unsigned (capable of representing only non-negative integers) or signed (capable of representing negative integers as well).
An integer value is typically specified in the source code of a program as a sequence of digits optionally prefixed with + or −. Some programming languages allow other notations, such as hexadecimal (base 16) or octal (base 8). Some programming languages also permit digit group separators.
The internal representation of this datum is the way the value is stored in the computer's memory. Unlike mathematical integers, a typical datum in a computer has some minimal and maximum possible value.
The most common representation of a positive integer is a string of bits, using the binary numeral system. The order of the memory bytes storing the bits varies; see endianness. The width, precision, or bitness of an integral type is the number of bits in its representation. An integral type with n bits can encode 2n numbers; for example an unsigned type typically represents the non-negative values 0 through 2n − 1. Other encodings of integer values to bit patterns are sometimes used, for example binary-coded decimal or Gray code, or as printed character codes such as ASCII.
There are four well-known ways to represent signed numbers in a binary computing system. The most common is two's complement, which allows a signed integral type with n bits to represent numbers from −2(n−1) through 2(n−1) − 1. Two's complement arithmetic is convenient because there is a perfect one-to-one correspondence between representations and values (in particular, no separate +0 and −0), and because addition, subtraction and multiplication do not need to distinguish between signed and unsigned types. Other possibilities include offset binary, sign-magnitude, and ones' complement.
Some computer languages define integer sizes in a machine-independent way; others have varying definitions depending on the underlying processor word size. Not all language implementations define variables of all integer sizes, and defined sizes may not even be distinct in a particular implementation. An integer in one programming language may be a different size in a different language, on a different processor, or in an execution context of different bitness; see § Words.
Some older computer architectures used decimal representations of integers, stored in binary-coded decimal (BCD) or other format. These values generally require data sizes of 4 bits per decimal digit (sometimes called a nibble), usually with additional bits for a sign. Many modern CPUs provide limited support for decimal integers as an extended datatype, providing instructions for converting such values to and from binary values. Depending on the architecture, decimal integers may have fixed sizes (e.g., 7 decimal digits plus a sign fit into a 32-bit word), or may be variable-length (up to some maximum digit size), typically occupying two digits per byte (octet).
Common integral data types
Bits | Name | Range (assuming two's complement for signed) | Decimal digits | Uses | Implementations | |||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
C/C++ | C# | Pascal and Delphi | Java | SQL | FORTRAN | D | Rust | |||||
4 | nibble, semioctet | Signed: From −8 to 7, from −(23) to 23 − 1 | 0.9 | Binary-coded decimal, single decimal digit representation | — | |||||||
Unsigned: From 0 to 15, which equals 24 − 1 | 1.2 | |||||||||||
8 | byte, octet, i8, u8 | Signed: From −128 to 127, from −(27) to 27 − 1 | 2.11 | ASCII characters, code units in the UTF-8 character encoding | int8_t, signed char | sbyte | Shortint | byte | tinyint | INTEGER | byte | i8 |
Unsigned: From 0 to 255, which equals 28 − 1 | 2.41 | uint8_t, unsigned char | byte | Byte | — | unsigned tinyint | — | ubyte | u8 | |||
16 | halfword, word, short, i16, u16 | Signed: From −32,768 to 32,767, from −(215) to 215 − 1 | 4.52 | UCS-2 characters, code units in the UTF-16 character encoding | int16_t, short,int | short | Smallint | short | smallint | INTEGER | short | i16 |
Unsigned: From 0 to 65,535, which equals 216 − 1 | 4.82 | uint16_t, unsigned,unsigned int | ushort | Word | char | unsigned smallint | — | ushort | u16 | |||
32 | word, long, doubleword, longword, int, i32, u32 | Signed: From −2,147,483,648 to 2,147,483,647, from −(231) to 231 − 1 | 9.33 | UTF-32 characters, true color with alpha, FourCC, pointers in 32-bit computing | int32_t, int,long | int | LongInt; Integer | int | int | INTEGER | int | i32 |
Unsigned: From 0 to 4,294,967,295, which equals 232 − 1 | 9.63 | uint32_t, unsigned,unsigned int,unsigned long | uint | LongWord; DWord; Cardinal | — | unsigned int | — | uint | u32 | |||
64 | word, doubleword, longword, long, long long, quad, quadword, qword, int64, i64, u64 | Signed: From −(263) to 263 − 1 | 18.96 | Time (e.g. milliseconds since the Unix epoch), pointers in 64-bit computing | int64_t, long,long long | long | Int64 | long | bigint | INTEGER | long | i64 |
Unsigned: From 0 to 264 − 1 | 19.27 | uint64_t, unsigned long long | ulong | UInt64; QWord | — | unsigned bigint | — | ulong | u64 | |||
128 | octaword, double quadword, i128, u128 | Signed: From −(2127) to 2127 − 1 | 38.23 | Complex scientific calculations, IPv6 addresses, GUIDs | Only available as non-standard or compiler-specific extensions | cent | i128 | |||||
Unsigned: From 0 to 2128 − 1 | 38.53 | ucent | u128 | |||||||||
n | n-bit integer (general case) | Signed: −(2n−1) to (2n−1 − 1) | (n − 1) log10 2 | C23: _BitInt(n), signed _BitInt(n) | Ada: range -2**(n-1)..2**(n-1)-1 | |||||||
Unsigned: 0 to (2n − 1) | n log10 2 | C23: unsigned _BitInt(n) | Ada: range 0..2**n-1 , mod 2**n ; standard libraries' or third-party arbitrary arithmetic libraries' BigDecimal or Decimal classes in many languages such as Python, C++, etc. |
Different CPUs support different integral data types. Typically, hardware will support both signed and unsigned types, but only a small, fixed set of widths.
The table above lists integral type widths that are supported in hardware by common processors. High-level programming languages provide more possibilities. It is common to have a 'double width' integral type that has twice as many bits as the biggest hardware-supported type. Many languages also have bit-field types (a specified number of bits, usually constrained to be less than the maximum hardware-supported width) and range types (that can represent only the integers in a specified range).
Some languages, such as Lisp, Smalltalk, REXX, Haskell, Python, and Raku, support arbitrary precision integers (also known as infinite precision integers or bignums). Other languages that do not support this concept as a top-level construct may have libraries available to represent very large numbers using arrays of smaller variables, such as Java's BigInteger class or Perl's "bigint" package. These use as much of the computer's memory as is necessary to store the numbers; however, a computer has only a finite amount of storage, so they, too, can only represent a finite subset of the mathematical integers. These schemes support very large numbers; for example one kilobyte of memory could be used to store numbers up to 2466 decimal digits long.
A Boolean type is a type that can represent only two values: 0 and 1, usually identified with false and true respectively. This type can be stored in memory using a single bit, but is often given a full byte for convenience of addressing and speed of access.
A four-bit quantity is known as a nibble (when eating, being smaller than a bite) or nybble (being a pun on the form of the word byte). One nibble corresponds to one digit in hexadecimal and holds one digit or a sign code in binary-coded decimal.
Bytes and octets
The term byte initially meant 'the smallest addressable unit of memory'. In the past, 5-, 6-, 7-, 8-, and 9-bit bytes have all been used. There have also been computers that could address individual bits ('bit-addressed machine'), or that could only address 16- or 32-bit quantities ('word-addressed machine'). The term byte was usually not used at all in connection with bit- and word-addressed machines.
The term octet always refers to an 8-bit quantity. It is mostly used in the field of computer networking, where computers with different byte widths might have to communicate.
In modern usage byte almost invariably means eight bits, since all other sizes have fallen into disuse; thus byte has come to be synonymous with octet.
Words
The term 'word' is used for a small group of bits that are handled simultaneously by processors of a particular architecture. The size of a word is thus CPU-specific. Many different word sizes have been used, including 6-, 8-, 12-, 16-, 18-, 24-, 32-, 36-, 39-, 40-, 48-, 60-, and 64-bit. Since it is architectural, the size of a word is usually set by the first CPU in a family, rather than the characteristics of a later compatible CPU. The meanings of terms derived from word, such as longword, doubleword, quadword, and halfword, also vary with the CPU and OS.
Practically all new desktop processors are capable of using 64-bit words, though embedded processors with 8- and 16-bit word size are still common. The 36-bit word length was common in the early days of computers.
One important cause of non-portability of software is the incorrect assumption that all computers have the same word size as the computer used by the programmer. For example, if a programmer using the C language incorrectly declares as int a variable that will be used to store values greater than 215−1, the program will fail on computers with 16-bit integers. That variable should have been declared as long, which has at least 32 bits on any computer. Programmers may also incorrectly assume that a pointer can be converted to an integer without loss of information, which may work on (some) 32-bit computers, but fail on 64-bit computers with 64-bit pointers and 32-bit integers. This issue is resolved by C99 in stdint.h in the form of intptr_t
.
The bitness of a program may refer to the word size (or bitness) of the processor on which it runs, or it may refer to the width of a memory address or pointer, which can differ between execution modes or contexts. For example, 64-bit versions of Microsoft Windows support existing 32-bit binaries, and programs compiled for Linux's x32 ABI run in 64-bit mode yet use 32-bit memory addresses.
Standard integer
The standard integer size is platform-dependent.
In C, it is denoted by int and required to be at least 16 bits. Windows and Unix systems have 32-bit ints on both 32-bit and 64-bit architectures.
Short integer
A short integer can represent a whole number that may take less storage, while having a smaller range, compared with a standard integer on the same machine.
In C, it is denoted by short. It is required to be at least 16 bits, and is often smaller than a standard integer, but this is not required. A conforming program can assume that it can safely store values between −(215−1) and 215−1, but it may not assume that the range is not larger. In Java, a short is always a 16-bit integer. In the Windows API, the datatype SHORT is defined as a 16-bit signed integer on all machines.
Programming language | Data type name | Signedness | Size in bytes | Minimum value | Maximum value |
---|---|---|---|---|---|
C and C++ | short | signed | 2 | −32,767 | +32,767 |
unsigned short | unsigned | 2 | 0 | 65,535 | |
C# | short | signed | 2 | −32,768 | +32,767 |
ushort | unsigned | 2 | 0 | 65,535 | |
Java | short | signed | 2 | −32,768 | +32,767 |
SQL | smallint | signed | 2 | −32,768 | +32,767 |
Long integer
A long integer can represent a whole integer whose range is greater than or equal to that of a standard integer on the same machine.
In C, it is denoted by long. It is required to be at least 32 bits, and may or may not be larger than a standard integer. A conforming program can assume that it can safely store values between −(231−1) and 231−1, but it may not assume that the range is not larger.
Programming language | Approval Type | Platforms | Data type name | Storage in bytes | Signed range | Unsigned range |
---|---|---|---|---|---|---|
C ISO/ANSI C99 | International Standard | Unix, 16/32-bit systems Windows, 16/32/64-bit systems | long | 4 (minimum requirement 4) | −2,147,483,647 to +2,147,483,647 | 0 to 4,294,967,295 (minimum requirement) |
C ISO/ANSI C99 | International Standard | Unix, 64-bit systems | long | 8 (minimum requirement 4) | −9,223,372,036,854,775,807 to +9,223,372,036,854,775,807 | 0 to 18,446,744,073,709,551,615 |
C++ ISO/ANSI | International Standard | Unix, Windows, 16/32-bit system | long | 4 (minimum requirement 4) | −2,147,483,648 to +2,147,483,647 | 0 to 4,294,967,295 (minimum requirement) |
C++/CLI | International Standard ECMA-372 | Unix, Windows, 16/32-bit systems | long | 4 (minimum requirement 4) | −2,147,483,648 to +2,147,483,647 | 0 to 4,294,967,295 (minimum requirement) |
VB | Company Standard | Windows | Long | 4 | −2,147,483,648 to +2,147,483,647 | — |
VBA | Company Standard | Windows, Mac OS X | Long | 4 | −2,147,483,648 to +2,147,483,647 | — |
SQL Server | Company Standard | Windows | BigInt | 8 | −9,223,372,036,854,775,808 to +9,223,372,036,854,775,807 | 0 to 18,446,744,073,709,551,615 |
C#/ VB.NET | ECMA International Standard | Microsoft .NET | long or Int64 | 8 | −9,223,372,036,854,775,808 to +9,223,372,036,854,775,807 | 0 to 18,446,744,073,709,551,615 |
Java | International/Company Standard | Java platform | long | 8 | −9,223,372,036,854,775,808 to +9,223,372,036,854,775,807 | — |
Pascal | ? | Windows, UNIX | int64 | 8 | −9,223,372,036,854,775,808 to +9,223,372,036,854,775,807 | 0 to 18,446,744,073,709,551,615 (Qword type) |
Long long
In the C99 version of the C programming language and the C++11 version of C++, a long long
type is supported that has double the minimum capacity of the standard long
. This type is not supported by compilers that require C code to be compliant with the previous C++ standard, C++03, because the long long type did not exist in C++03. For an ANSI/ISO compliant compiler, the minimum requirements for the specified ranges, that is, −(263−1) to 263−1 for signed and 0 to 264−1 for unsigned, must be fulfilled; however, extending this range is permitted. This can be an issue when exchanging code and data between platforms, or doing direct hardware access. Thus, there are several sets of headers providing platform independent exact width types. The C standard library provides stdint.h; this was introduced in C99 and C++11.
Syntax
Integer literals can be written as regular Arabic numerals, consisting of a sequence of digits and with negation indicated by a minus sign before the value. However, most programming languages disallow use of commas or spaces for digit grouping. Examples of integer literals are:
42
10000
-233000
There are several alternate methods for writing integer literals in many programming languages:
- Many programming languages, especially those influenced by C, prefix an integer literal with
0X
or0x
to represent a hexadecimal value, e.g.0xDEADBEEF
. Other languages may use a different notation, e.g. some assembly languages append anH
orh
to the end of a hexadecimal value. - Perl, Ruby, Java, Julia, D, Go, C#, Rust and Python (starting from version 3.6) allow embedded underscores for clarity, e.g.
10_000_000
, and fixed-form Fortran ignores embedded spaces in integer literals. C (starting from C23) and C++ use single quotes for this purpose. - In C and C++, a leading zero indicates an octal value, e.g.
0755
. This was primarily intended to be used with Unix modes; however, it has been criticized because normal integers may also lead with zero. As such, Python, Ruby, Haskell, and OCaml prefix octal values with0O
or0o
, following the layout used by hexadecimal values. - Several languages, including Java, C#, Scala, Python, Ruby, OCaml, C (starting from C23) and C++ can represent binary values by prefixing a number with
0B
or0b
.
Extreme values
In many programming languages, there exist predefined constants representing the least and the greatest values representable with a given integer type.
Names for these include
- SmallBASIC:
MAXINT
- Java:
java.lang.Integer.MAX_VALUE
,java.lang.Integer.MIN_VALUE
- Corresponding fields exist for the other integer classes in Java.
- C:
INT_MAX
, etc.[citation needed]- GLib:
G_MININT
,G_MAXINT
,G_MAXUINT
, ...
- GLib:
- Pascal:
MaxInt
[citation needed] - Python 2:
sys.maxint
[citation needed] - Turing:
maxint
See also
- Arbitrary-precision arithmetic
- Binary-coded decimal (BCD)
- C data types
- Integer overflow
- Signed number representations
Notes
- Not all SQL dialects have unsigned datatypes.
- The sizes of char, short, int, long and long long in C/C++ are dependent upon the implementation of the language.
- Fortan uses 'kinds' to control the size of integers. Parameterized constants defining the available kinds are available in the iso_fortran_env intrinsic module. Constants defining C compatible kinds are available in the iso_c_binding intrinsic module.
- Java does not directly support arithmetic on char types. The results must be cast back into char from an int.
- The sizes of Delphi's Integer and Cardinal are not guaranteed, varying from platform to platform; usually defined as LongInt and LongWord respectively.
- Reserved for future use. Not implemented yet.
- The ISO C standard allows implementations to reserve the value with sign bit 1 and all other bits 0 (for sign–magnitude and two's complement representation) or with all bits 1 (for ones' complement) for use as a "trap" value, used to indicate (for example) an overflow.
References
- Cheever, Eric. "Representation of numbers". Swarthmore College. Retrieved 2011-09-11.
- Madhusudhan Konda (2011-09-02). "A look at Java 7's new features - O'Reilly Radar". Radar.oreilly.com. Retrieved 2013-10-15.
- Barr, Adam (2018-10-23). The Problem with Software: Why Smart Engineers Write Bad Code. MIT Press. ISBN 978-0-262-34821-8.
- "Sybase Adaptive Server Enterprise 15.5: Exact Numeric Datatypes".
- "MySQL 5.6 Numeric Datatypes".
- "BigInteger (Java Platform SE 6)". Oracle. Retrieved 2011-09-11.
- Fog, Agner (2010-02-16). "Calling conventions for different C++ compilers and operating systems: Chapter 3, Data Representation" (PDF). Retrieved 2010-08-30.
- Thorsten Leemhuis (2011-09-13). "Kernel Log: x32 ABI gets around 64-bit drawbacks". www.h-online.com. Archived from the original on 28 October 2011. Retrieved 2011-11-01.
- Giguere, Eric (1987-12-18). "The ANSI Standard: A Summary for the C Programmer". Retrieved 2010-09-04.
- Meyers, Randy (2000-12-01). "The New C: Integers in C99, Part 1". drdobbs.com. Retrieved 2010-09-04.
- "ISO/IEC 9899:201x" (PDF). open-std.org. section 6.2.6.2, paragraph 2. Retrieved 2016-06-20.
- "ISO/IEC 9899:201x" (PDF). open-std.org. section 5.2.4.2.1. Retrieved 2016-06-20.
- "Fundamental types in C++". cppreference.com. Retrieved 5 December 2010.
- "Chapter 8.6.2 on page 12" (PDF). ecma-international.org.
- VB 6.0 help file
- "The Integer, Long, and Byte Data Types (VBA)". microsoft.com. Retrieved 2006-12-19.
- Giguere, Eric (December 18, 1987). "The ANSI Standard: A Summary for the C Programmer". Retrieved 2010-09-04.
- "American National Standard Programming Language C specifies the syntax and semantics of programs written in the C programming language". Archived from the original on 2010-08-22. Retrieved 2010-09-04.
- ECMAScript 6th Edition draft: https://people.mozilla.org/~jorendorff/es6-draft.html#sec-literals-numeric-literals Archived 2013-12-16 at the Wayback Machine
- "SmallBASIC | MAXINT". Retrieved 2025-01-20.
- "Integer (Java Platform SE 8 )". Retrieved 2025-01-20.
- "Limits of Basic Types". Retrieved 2025-01-20.
- Grogono, Peter (1995). Programming with Turing and Object Oriented Turing. New York: Springer. p. 363. doi:10.1007/978-1-4612-4238-3. LCCN 95010802.
In computer science an integer is a datum of integral data type a data type that represents some range of mathematical integers Integral data types may be of different sizes and may or may not be allowed to contain negative values Integers are commonly represented in a computer as a group of binary digits bits The size of the grouping varies so the set of integer sizes available varies between different types of computers Computer hardware nearly always provides a way to represent a processor register or memory address as an integer Value and representationThe value of an item with an integral type is the mathematical integer that it corresponds to Integral types may be unsigned capable of representing only non negative integers or signed capable of representing negative integers as well An integer value is typically specified in the source code of a program as a sequence of digits optionally prefixed with or Some programming languages allow other notations such as hexadecimal base 16 or octal base 8 Some programming languages also permit digit group separators The internal representation of this datum is the way the value is stored in the computer s memory Unlike mathematical integers a typical datum in a computer has some minimal and maximum possible value The most common representation of a positive integer is a string of bits using the binary numeral system The order of the memory bytes storing the bits varies see endianness The width precision or bitness of an integral type is the number of bits in its representation An integral type with n bits can encode 2n numbers for example an unsigned type typically represents the non negative values 0 through 2n 1 Other encodings of integer values to bit patterns are sometimes used for example binary coded decimal or Gray code or as printed character codes such as ASCII There are four well known ways to represent signed numbers in a binary computing system The most common is two s complement which allows a signed integral type with n bits to represent numbers from 2 n 1 through 2 n 1 1 Two s complement arithmetic is convenient because there is a perfect one to one correspondence between representations and values in particular no separate 0 and 0 and because addition subtraction and multiplication do not need to distinguish between signed and unsigned types Other possibilities include offset binary sign magnitude and ones complement Some computer languages define integer sizes in a machine independent way others have varying definitions depending on the underlying processor word size Not all language implementations define variables of all integer sizes and defined sizes may not even be distinct in a particular implementation An integer in one programming language may be a different size in a different language on a different processor or in an execution context of different bitness see Words Some older computer architectures used decimal representations of integers stored in binary coded decimal BCD or other format These values generally require data sizes of 4 bits per decimal digit sometimes called a nibble usually with additional bits for a sign Many modern CPUs provide limited support for decimal integers as an extended datatype providing instructions for converting such values to and from binary values Depending on the architecture decimal integers may have fixed sizes e g 7 decimal digits plus a sign fit into a 32 bit word or may be variable length up to some maximum digit size typically occupying two digits per byte octet Common integral data typesBits Name Range assuming two s complement for signed Decimal digits Uses ImplementationsC C C Pascal and Delphi Java SQL FORTRAN D Rust4 nibble semioctet Signed From 8 to 7 from 23 to 23 1 0 9 Binary coded decimal single decimal digit repre sen ta tion Unsigned From 0 to 15 which equals 24 1 1 28 byte octet i8 u8 Signed From 128 to 127 from 27 to 27 1 2 11 ASCII characters code units in the UTF 8 character encoding int8 t signed char sbyte Shortint byte tinyint INTEGER byte i8Unsigned From 0 to 255 which equals 28 1 2 41 uint8 t unsigned char byte Byte unsigned tinyint ubyte u816 halfword word short i16 u16 Signed From 32 768 to 32 767 from 215 to 215 1 4 52 UCS 2 characters code units in the UTF 16 character encoding int16 t short int short Smallint short smallint INTEGER short i16Unsigned From 0 to 65 535 which equals 216 1 4 82 uint16 t unsigned unsigned int ushort Word char unsigned smallint ushort u1632 word long doubleword longword int i32 u32 Signed From 2 147 483 648 to 2 147 483 647 from 231 to 231 1 9 33 UTF 32 characters true color with alpha FourCC pointers in 32 bit computing int32 t int long int LongInt Integer int int INTEGER int i32Unsigned From 0 to 4 294 967 295 which equals 232 1 9 63 uint32 t unsigned unsigned int unsigned long uint LongWord DWord Cardinal unsigned int uint u3264 word doubleword longword long long long quad quadword qword int64 i64 u64 Signed From 263 to 263 1 18 96 Time e g milli seconds since the Unix epoch pointers in 64 bit computing int64 t long long long long Int64 long bigint INTEGER long i64Unsigned From 0 to 264 1 19 27 uint64 t unsigned long long ulong UInt64 QWord unsigned bigint ulong u64128 octaword double quadword i128 u128 Signed From 2127 to 2127 1 38 23 Complex scientific cal cula tions IPv6 addresses GUIDs Only available as non standard or compiler specific extensions cent i128Unsigned From 0 to 2128 1 38 53 ucent u128n n bit integer general case Signed 2n 1 to 2n 1 1 n 1 log10 2 C23 BitInt n signed BitInt n Ada span class kr range span span class o span span class mi 2 span span class o span span class p span span class n n span span class o span span class mi 1 span span class p span span class mi 2 span span class o span span class p span span class n n span span class o span span class mi 1 span span class p span span class o span span class mi 1 span Unsigned 0 to 2n 1 n log10 2 C23 unsigned BitInt n Ada span class kr range span span class mf 0 span span class p span span class mi 2 span span class o span span class n n span span class o span span class mi 1 span span class ow mod span span class mi 2 span span class o span span class n n span standard libraries or third party arbitrary arithmetic libraries BigDecimal or Decimal classes in many languages such as Python C etc Different CPUs support different integral data types Typically hardware will support both signed and unsigned types but only a small fixed set of widths The table above lists integral type widths that are supported in hardware by common processors High level programming languages provide more possibilities It is common to have a double width integral type that has twice as many bits as the biggest hardware supported type Many languages also have bit field types a specified number of bits usually constrained to be less than the maximum hardware supported width and range types that can represent only the integers in a specified range Some languages such as Lisp Smalltalk REXX Haskell Python and Raku support arbitrary precision integers also known as infinite precision integers or bignums Other languages that do not support this concept as a top level construct may have libraries available to represent very large numbers using arrays of smaller variables such as Java s BigInteger class or Perl s bigint package These use as much of the computer s memory as is necessary to store the numbers however a computer has only a finite amount of storage so they too can only represent a finite subset of the mathematical integers These schemes support very large numbers for example one kilobyte of memory could be used to store numbers up to 2466 decimal digits long A Boolean type is a type that can represent only two values 0 and 1 usually identified with false and true respectively This type can be stored in memory using a single bit but is often given a full byte for convenience of addressing and speed of access A four bit quantity is known as a nibble when eating being smaller than a bite or nybble being a pun on the form of the word byte One nibble corresponds to one digit in hexadecimal and holds one digit or a sign code in binary coded decimal Bytes and octets The term byte initially meant the smallest addressable unit of memory In the past 5 6 7 8 and 9 bit bytes have all been used There have also been computers that could address individual bits bit addressed machine or that could only address 16 or 32 bit quantities word addressed machine The term byte was usually not used at all in connection with bit and word addressed machines The term octet always refers to an 8 bit quantity It is mostly used in the field of computer networking where computers with different byte widths might have to communicate In modern usage byte almost invariably means eight bits since all other sizes have fallen into disuse thus byte has come to be synonymous with octet Words The term word is used for a small group of bits that are handled simultaneously by processors of a particular architecture The size of a word is thus CPU specific Many different word sizes have been used including 6 8 12 16 18 24 32 36 39 40 48 60 and 64 bit Since it is architectural the size of a word is usually set by the first CPU in a family rather than the characteristics of a later compatible CPU The meanings of terms derived from word such as longword doubleword quadword and halfword also vary with the CPU and OS Practically all new desktop processors are capable of using 64 bit words though embedded processors with 8 and 16 bit word size are still common The 36 bit word length was common in the early days of computers One important cause of non portability of software is the incorrect assumption that all computers have the same word size as the computer used by the programmer For example if a programmer using the C language incorrectly declares as int a variable that will be used to store values greater than 215 1 the program will fail on computers with 16 bit integers That variable should have been declared as long which has at least 32 bits on any computer Programmers may also incorrectly assume that a pointer can be converted to an integer without loss of information which may work on some 32 bit computers but fail on 64 bit computers with 64 bit pointers and 32 bit integers This issue is resolved by C99 in stdint h in the form of intptr t The bitness of a program may refer to the word size or bitness of the processor on which it runs or it may refer to the width of a memory address or pointer which can differ between execution modes or contexts For example 64 bit versions of Microsoft Windows support existing 32 bit binaries and programs compiled for Linux s x32 ABI run in 64 bit mode yet use 32 bit memory addresses Standard integer The standard integer size is platform dependent In C it is denoted by int and required to be at least 16 bits Windows and Unix systems have 32 bit int s on both 32 bit and 64 bit architectures Short integer A short integer can represent a whole number that may take less storage while having a smaller range compared with a standard integer on the same machine In C it is denoted by short It is required to be at least 16 bits and is often smaller than a standard integer but this is not required A conforming program can assume that it can safely store values between 215 1 and 215 1 but it may not assume that the range is not larger In Java a short is always a 16 bit integer In the Windows API the datatype SHORT is defined as a 16 bit signed integer on all machines Common short integer sizes Programming language Data type name Signedness Size in bytes Minimum value Maximum valueC and C short signed 2 32 767 32 767unsigned short unsigned 2 0 65 535C short signed 2 32 768 32 767ushort unsigned 2 0 65 535Java short signed 2 32 768 32 767SQL smallint signed 2 32 768 32 767Long integer A long integer can represent a whole integer whose range is greater than or equal to that of a standard integer on the same machine In C it is denoted by long It is required to be at least 32 bits and may or may not be larger than a standard integer A conforming program can assume that it can safely store values between 231 1 and 231 1 but it may not assume that the range is not larger Common long integer sizes Programming language Approval Type Platforms Data type name Storage in bytes Signed range Unsigned rangeC ISO ANSI C99 International Standard Unix 16 32 bit systems Windows 16 32 64 bit systems long 4 minimum require ment 4 2 147 483 647 to 2 147 483 647 0 to 4 294 967 295 minimum require ment C ISO ANSI C99 International Standard Unix 64 bit systems long 8 minimum require ment 4 9 223 372 036 854 775 807 to 9 223 372 036 854 775 807 0 to 18 446 744 073 709 551 615C ISO ANSI International Standard Unix Windows 16 32 bit system long 4 minimum require ment 4 2 147 483 648 to 2 147 483 647 0 to 4 294 967 295 minimum require ment C CLI International Standard ECMA 372 Unix Windows 16 32 bit systems long 4 minimum require ment 4 2 147 483 648 to 2 147 483 647 0 to 4 294 967 295 minimum require ment VB Company Standard Windows Long 4 2 147 483 648 to 2 147 483 647 VBA Company Standard Windows Mac OS X Long 4 2 147 483 648 to 2 147 483 647 SQL Server Company Standard Windows BigInt 8 9 223 372 036 854 775 808 to 9 223 372 036 854 775 807 0 to 18 446 744 073 709 551 615C VB NET ECMA International Standard Microsoft NET long or Int64 8 9 223 372 036 854 775 808 to 9 223 372 036 854 775 807 0 to 18 446 744 073 709 551 615Java International Company Standard Java platform long 8 9 223 372 036 854 775 808 to 9 223 372 036 854 775 807 Pascal Windows UNIX int64 8 9 223 372 036 854 775 808 to 9 223 372 036 854 775 807 0 to 18 446 744 073 709 551 615 Qword type Long long In the C99 version of the C programming language and the C 11 version of C a long long type is supported that has double the minimum capacity of the standard long This type is not supported by compilers that require C code to be compliant with the previous C standard C 03 because the long long type did not exist in C 03 For an ANSI ISO compliant compiler the minimum requirements for the specified ranges that is 263 1 to 263 1 for signed and 0 to 264 1 for unsigned must be fulfilled however extending this range is permitted This can be an issue when exchanging code and data between platforms or doing direct hardware access Thus there are several sets of headers providing platform independent exact width types The C standard library provides stdint h this was introduced in C99 and C 11 SyntaxInteger literals can be written as regular Arabic numerals consisting of a sequence of digits and with negation indicated by a minus sign before the value However most programming languages disallow use of commas or spaces for digit grouping Examples of integer literals are 42 10000 233000 There are several alternate methods for writing integer literals in many programming languages Many programming languages especially those influenced by C prefix an integer literal with 0X or 0x to represent a hexadecimal value e g 0xDEADBEEF Other languages may use a different notation e g some assembly languages append an H or h to the end of a hexadecimal value Perl Ruby Java Julia D Go C Rust and Python starting from version 3 6 allow embedded underscores for clarity e g 10 000 000 and fixed form Fortran ignores embedded spaces in integer literals C starting from C23 and C use single quotes for this purpose In C and C a leading zero indicates an octal value e g 0755 This was primarily intended to be used with Unix modes however it has been criticized because normal integers may also lead with zero As such Python Ruby Haskell and OCaml prefix octal values with 0O or 0o following the layout used by hexadecimal values Several languages including Java C Scala Python Ruby OCaml C starting from C23 and C can represent binary values by prefixing a number with 0B or 0b Extreme valuesIn many programming languages there exist predefined constants representing the least and the greatest values representable with a given integer type Names for these include SmallBASIC span class vg MAXINT span Java span class n java span span class p span span class na lang span span class p span span class na Integer span span class p span span class na MAX VALUE span span class n java span span class p span span class na lang span span class p span span class na Integer span span class p span span class na MIN VALUE span Corresponding fields exist for the other integer classes in Java C span class n INT MAX span etc citation needed GLib span class n G MININT span span class n G MAXINT span span class n G MAXUINT span Pascal span class n MaxInt span citation needed Python 2 span class n sys span span class o span span class n maxint span citation needed Turing maxintSee alsoArbitrary precision arithmetic Binary coded decimal BCD C data types Integer overflow Signed number representationsNotesNot all SQL dialects have unsigned datatypes The sizes of char short int long and long long in C C are dependent upon the implementation of the language Fortan uses kinds to control the size of integers Parameterized constants defining the available kinds are available in the iso fortran env intrinsic module Constants defining C compatible kinds are available in the iso c binding intrinsic module Java does not directly support arithmetic on char types The results must be cast back into char from an int The sizes of Delphi s Integer and Cardinal are not guaranteed varying from platform to platform usually defined as LongInt and LongWord respectively Reserved for future use Not implemented yet The ISO C standard allows implementations to reserve the value with sign bit 1 and all other bits 0 for sign magnitude and two s complement representation or with all bits 1 for ones complement for use as a trap value used to indicate for example an overflow ReferencesCheever Eric Representation of numbers Swarthmore College Retrieved 2011 09 11 Madhusudhan Konda 2011 09 02 A look at Java 7 s new features O Reilly Radar Radar oreilly com Retrieved 2013 10 15 Barr Adam 2018 10 23 The Problem with Software Why Smart Engineers Write Bad Code MIT Press ISBN 978 0 262 34821 8 Sybase Adaptive Server Enterprise 15 5 Exact Numeric Datatypes MySQL 5 6 Numeric Datatypes BigInteger Java Platform SE 6 Oracle Retrieved 2011 09 11 Fog Agner 2010 02 16 Calling conventions for different C compilers and operating systems Chapter 3 Data Representation PDF Retrieved 2010 08 30 Thorsten Leemhuis 2011 09 13 Kernel Log x32 ABI gets around 64 bit drawbacks www h online com Archived from the original on 28 October 2011 Retrieved 2011 11 01 Giguere Eric 1987 12 18 The ANSI Standard A Summary for the C Programmer Retrieved 2010 09 04 Meyers Randy 2000 12 01 The New C Integers in C99 Part 1 drdobbs com Retrieved 2010 09 04 ISO IEC 9899 201x PDF open std org section 6 2 6 2 paragraph 2 Retrieved 2016 06 20 ISO IEC 9899 201x PDF open std org section 5 2 4 2 1 Retrieved 2016 06 20 Fundamental types in C cppreference com Retrieved 5 December 2010 Chapter 8 6 2 on page 12 PDF ecma international org VB 6 0 help file The Integer Long and Byte Data Types VBA microsoft com Retrieved 2006 12 19 Giguere Eric December 18 1987 The ANSI Standard A Summary for the C Programmer Retrieved 2010 09 04 American National Standard Programming Language C specifies the syntax and semantics of programs written in the C programming language Archived from the original on 2010 08 22 Retrieved 2010 09 04 ECMAScript 6th Edition draft https people mozilla org jorendorff es6 draft html sec literals numeric literals Archived 2013 12 16 at the Wayback Machine SmallBASIC MAXINT Retrieved 2025 01 20 Integer Java Platform SE 8 Retrieved 2025 01 20 Limits of Basic Types Retrieved 2025 01 20 Grogono Peter 1995 Programming with Turing and Object Oriented Turing New York Springer p 363 doi 10 1007 978 1 4612 4238 3 LCCN 95010802