Word bytes and bits

From Wikipedia, the free encyclopedia

In computing, a word is the natural unit of data used by a particular processor design. A word is a fixed-sized datum handled as a unit by the instruction set or the hardware of the processor. The number of bits or digits[a] in a word (the word size, word width, or word length) is an important characteristic of any specific processor design or computer architecture.

The size of a word is reflected in many aspects of a computer’s structure and operation; the majority of the registers in a processor are usually word-sized and the largest datum that can be transferred to and from the working memory in a single operation is a word in many (not all) architectures. The largest possible address size, used to designate a location in memory, is typically a hardware word (here, «hardware word» means the full-sized natural word of the processor, as opposed to any other definition used).

Documentation for older computers with fixed word size commonly states memory sizes in words rather than bytes or characters. The documentation sometimes uses metric prefixes correctly, sometimes with rounding, e.g., 65 kilowords (KW) meaning for 65536 words, and sometimes uses them incorrectly, with kilowords (KW) meaning 1024 words (210) and megawords (MW) meaning 1,048,576 words (220). With standardization on 8-bit bytes and byte addressability, stating memory sizes in bytes, kilobytes, and megabytes with powers of 1024 rather than 1000 has become the norm, although there is some use of the IEC binary prefixes.

Several of the earliest computers (and a few modern as well) use binary-coded decimal rather than plain binary, typically having a word size of 10 or 12 decimal digits, and some early decimal computers have no fixed word length at all. Early binary systems tended to use word lengths that were some multiple of 6-bits, with the 36-bit word being especially common on mainframe computers. The introduction of ASCII led to the move to systems with word lengths that were a multiple of 8-bits, with 16-bit machines being popular in the 1970s before the move to modern processors with 32 or 64 bits.[1] Special-purpose designs like digital signal processors, may have any word length from 4 to 80 bits.[1]

The size of a word can sometimes differ from the expected due to backward compatibility with earlier computers. If multiple compatible variations or a family of processors share a common architecture and instruction set but differ in their word sizes, their documentation and software may become notationally complex to accommodate the difference (see Size families below).

Uses of words[edit]

Depending on how a computer is organized, word-size units may be used for:

Fixed-point numbers
Holders for fixed point, usually integer, numerical values may be available in one or in several different sizes, but one of the sizes available will almost always be the word. The other sizes, if any, are likely to be multiples or fractions of the word size. The smaller sizes are normally used only for efficient use of memory; when loaded into the processor, their values usually go into a larger, word sized holder.
Floating-point numbers
Holders for floating-point numerical values are typically either a word or a multiple of a word.
Addresses
Holders for memory addresses must be of a size capable of expressing the needed range of values but not be excessively large, so often the size used is the word though it can also be a multiple or fraction of the word size.
Registers
Processor registers are designed with a size appropriate for the type of data they hold, e.g. integers, floating-point numbers, or addresses. Many computer architectures use general-purpose registers that are capable of storing data in multiple representations.
Memory–processor transfer
When the processor reads from the memory subsystem into a register or writes a register’s value to memory, the amount of data transferred is often a word. Historically, this amount of bits which could be transferred in one cycle was also called a catena in some environments (such as the Bull GAMMA 60 [fr]).[2][3] In simple memory subsystems, the word is transferred over the memory data bus, which typically has a width of a word or half-word. In memory subsystems that use caches, the word-sized transfer is the one between the processor and the first level of cache; at lower levels of the memory hierarchy larger transfers (which are a multiple of the word size) are normally used.
Unit of address resolution
In a given architecture, successive address values designate successive units of memory; this unit is the unit of address resolution. In most computers, the unit is either a character (e.g. a byte) or a word. (A few computers have used bit resolution.) If the unit is a word, then a larger amount of memory can be accessed using an address of a given size at the cost of added complexity to access individual characters. On the other hand, if the unit is a byte, then individual characters can be addressed (i.e. selected during the memory operation).
Instructions
Machine instructions are normally the size of the architecture’s word, such as in RISC architectures, or a multiple of the «char» size that is a fraction of it. This is a natural choice since instructions and data usually share the same memory subsystem. In Harvard architectures the word sizes of instructions and data need not be related, as instructions and data are stored in different memories; for example, the processor in the 1ESS electronic telephone switch has 37-bit instructions and 23-bit data words.

Word size choice[edit]

When a computer architecture is designed, the choice of a word size is of substantial importance. There are design considerations which encourage particular bit-group sizes for particular uses (e.g. for addresses), and these considerations point to different sizes for different uses. However, considerations of economy in design strongly push for one size, or a very few sizes related by multiples or fractions (submultiples) to a primary size. That preferred size becomes the word size of the architecture.

Character size was in the past (pre-variable-sized character encoding) one of the influences on unit of address resolution and the choice of word size. Before the mid-1960s, characters were most often stored in six bits; this allowed no more than 64 characters, so the alphabet was limited to upper case. Since it is efficient in time and space to have the word size be a multiple of the character size, word sizes in this period were usually multiples of 6 bits (in binary machines). A common choice then was the 36-bit word, which is also a good size for the numeric properties of a floating point format.

After the introduction of the IBM System/360 design, which uses eight-bit characters and supports lower-case letters, the standard size of a character (or more accurately, a byte) becomes eight bits. Word sizes thereafter are naturally multiples of eight bits, with 16, 32, and 64 bits being commonly used.

Variable-word architectures[edit]

Early machine designs included some that used what is often termed a variable word length. In this type of organization, an operand has no fixed length. Depending on the machine and the instruction, the length might be denoted by a count field, by a delimiting character, or by an additional bit called, e.g., flag, or word mark. Such machines often use binary-coded decimal in 4-bit digits, or in 6-bit characters, for numbers. This class of machines includes the IBM 702, IBM 705, IBM 7080, IBM 7010, UNIVAC 1050, IBM 1401, IBM 1620, and RCA 301.

Most of these machines work on one unit of memory at a time and since each instruction or datum is several units long, each instruction takes several cycles just to access memory. These machines are often quite slow because of this. For example, instruction fetches on an IBM 1620 Model I take 8 cycles (160 μs) just to read the 12 digits of the instruction (the Model II reduced this to 6 cycles, or 4 cycles if the instruction did not need both address fields). Instruction execution takes a variable number of cycles, depending on the size of the operands.

Word, bit and byte addressing[edit]

The memory model of an architecture is strongly influenced by the word size. In particular, the resolution of a memory address, that is, the smallest unit that can be designated by an address, has often been chosen to be the word. In this approach, the word-addressable machine approach, address values which differ by one designate adjacent memory words. This is natural in machines which deal almost always in word (or multiple-word) units, and has the advantage of allowing instructions to use minimally sized fields to contain addresses, which can permit a smaller instruction size or a larger variety of instructions.

When byte processing is to be a significant part of the workload, it is usually more advantageous to use the byte, rather than the word, as the unit of address resolution. Address values which differ by one designate adjacent bytes in memory. This allows an arbitrary character within a character string to be addressed straightforwardly. A word can still be addressed, but the address to be used requires a few more bits than the word-resolution alternative. The word size needs to be an integer multiple of the character size in this organization. This addressing approach was used in the IBM 360, and has been the most common approach in machines designed since then.

When the workload involves processing fields of different sizes, it can be advantageous to address to the bit. Machines with bit addressing may have some instructions that use a programmer-defined byte size and other instructions that operate on fixed data sizes. As an example, on the IBM 7030[4] («Stretch»), a floating point instruction can only address words while an integer arithmetic instruction can specify a field length of 1-64 bits, a byte size of 1-8 bits and an accumulator offset of 0-127 bits.

In a byte-addressable machine with storage-to-storage (SS) instructions, there are typically move instructions to copy one or multiple bytes from one arbitrary location to another. In a byte-oriented (byte-addressable) machine without SS instructions, moving a single byte from one arbitrary location to another is typically:

  1. LOAD the source byte
  2. STORE the result back in the target byte

Individual bytes can be accessed on a word-oriented machine in one of two ways. Bytes can be manipulated by a combination of shift and mask operations in registers. Moving a single byte from one arbitrary location to another may require the equivalent of the following:

  1. LOAD the word containing the source byte
  2. SHIFT the source word to align the desired byte to the correct position in the target word
  3. AND the source word with a mask to zero out all but the desired bits
  4. LOAD the word containing the target byte
  5. AND the target word with a mask to zero out the target byte
  6. OR the registers containing the source and target words to insert the source byte
  7. STORE the result back in the target location

Alternatively many word-oriented machines implement byte operations with instructions using special byte pointers in registers or memory. For example, the PDP-10 byte pointer contained the size of the byte in bits (allowing different-sized bytes to be accessed), the bit position of the byte within the word, and the word address of the data. Instructions could automatically adjust the pointer to the next byte on, for example, load and deposit (store) operations.

Powers of two[edit]

Different amounts of memory are used to store data values with different degrees of precision. The commonly used sizes are usually a power of two multiple of the unit of address resolution (byte or word). Converting the index of an item in an array into the memory address offset of the item then requires only a shift operation rather than a multiplication. In some cases this relationship can also avoid the use of division operations. As a result, most modern computer designs have word sizes (and other operand sizes) that are a power of two times the size of a byte.

Size families[edit]

As computer designs have grown more complex, the central importance of a single word size to an architecture has decreased. Although more capable hardware can use a wider variety of sizes of data, market forces exert pressure to maintain backward compatibility while extending processor capability. As a result, what might have been the central word size in a fresh design has to coexist as an alternative size to the original word size in a backward compatible design. The original word size remains available in future designs, forming the basis of a size family.

In the mid-1970s, DEC designed the VAX to be a 32-bit successor of the 16-bit PDP-11. They used word for a 16-bit quantity, while longword referred to a 32-bit quantity; this terminology is the same as the terminology used for the PDP-11. This was in contrast to earlier machines, where the natural unit of addressing memory would be called a word, while a quantity that is one half a word would be called a halfword. In fitting with this scheme, a VAX quadword is 64 bits. They continued this 16-bit word/32-bit longword/64-bit quadword terminology with the 64-bit Alpha.

Another example is the x86 family, of which processors of three different word lengths (16-bit, later 32- and 64-bit) have been released, while word continues to designate a 16-bit quantity. As software is routinely ported from one word-length to the next, some APIs and documentation define or refer to an older (and thus shorter) word-length than the full word length on the CPU that software may be compiled for. Also, similar to how bytes are used for small numbers in many programs, a shorter word (16 or 32 bits) may be used in contexts where the range of a wider word is not needed (especially where this can save considerable stack space or cache memory space). For example, Microsoft’s Windows API maintains the programming language definition of WORD as 16 bits, despite the fact that the API may be used on a 32- or 64-bit x86 processor, where the standard word size would be 32 or 64 bits, respectively. Data structures containing such different sized words refer to them as:

  • WORD (16 bits/2 bytes)
  • DWORD (32 bits/4 bytes)
  • QWORD (64 bits/8 bytes)

A similar phenomenon has developed in Intel’s x86 assembly language – because of the support for various sizes (and backward compatibility) in the instruction set, some instruction mnemonics carry «d» or «q» identifiers denoting «double-«, «quad-» or «double-quad-«, which are in terms of the architecture’s original 16-bit word size.

An example with a different word size is the IBM System/360 family. In the System/360 architecture, System/370 architecture and System/390 architecture, there are 8-bit bytes, 16-bit halfwords, 32-bit words and 64-bit doublewords. The z/Architecture, which is the 64-bit member of that architecture family, continues to refer to 16-bit halfwords, 32-bit words, and 64-bit doublewords, and additionally features 128-bit quadwords.

In general, new processors must use the same data word lengths and virtual address widths as an older processor to have binary compatibility with that older processor.

Often carefully written source code – written with source-code compatibility and software portability in mind – can be recompiled to run on a variety of processors, even ones with different data word lengths or different address widths or both.

Table of word sizes[edit]

key: bit: bits, c: characters, d: decimal digits, w: word size of architecture, n: variable size, wm: Word mark
Year Computer
architecture
Word size w Integer
sizes
Floating­point
sizes
Instruction
sizes
Unit of address
resolution
Char size
1837 Babbage
Analytical engine
50 d w Five different cards were used for different functions, exact size of cards not known. w
1941 Zuse Z3 22 bit w 8 bit w
1942 ABC 50 bit w
1944 Harvard Mark I 23 d w 24 bit
1946
(1948)
{1953}
ENIAC
(w/Panel #16[5])
{w/Panel #26[6]}
10 d w, 2w
(w)
{w}

(2 d, 4 d, 6 d, 8 d)
{2 d, 4 d, 6 d, 8 d}


{w}
1948 Manchester Baby 32 bit w w w
1951 UNIVAC I 12 d w 12w w 1 d
1952 IAS machine 40 bit w 12w w 5 bit
1952 Fast Universal Digital Computer M-2 34 bit w? w 34 bit = 4-bit opcode plus 3×10 bit address 10 bit
1952 IBM 701 36 bit 12w, w 12w 12w, w 6 bit
1952 UNIVAC 60 n d 1 d, … 10 d 2 d, 3 d
1952 ARRA I 30 bit w w w 5 bit
1953 IBM 702 n c 0 c, … 511 c 5 c c 6 bit
1953 UNIVAC 120 n d 1 d, … 10 d 2 d, 3 d
1953 ARRA II 30 bit w 2w 12w w 5 bit
1954
(1955)
IBM 650
(w/IBM 653)
10 d w
(w)
w w 2 d
1954 IBM 704 36 bit w w w w 6 bit
1954 IBM 705 n c 0 c, … 255 c 5 c c 6 bit
1954 IBM NORC 16 d w w, 2w w w
1956 IBM 305 n d 1 d, … 100 d 10 d d 1 d
1956 ARMAC 34 bit w w 12w w 5 bit, 6 bit
1956 LGP-30 31 bit w 16 bit w 6 bit
1957 Autonetics Recomp I 40 bit w, 79 bit, 8 d, 15 d 12w 12w, w 5 bit
1958 UNIVAC II 12 d w 12w w 1 d
1958 SAGE 32 bit 12w w w 6 bit
1958 Autonetics Recomp II 40 bit w, 79 bit, 8 d, 15 d 2w 12w 12w, w 5 bit
1958 Setun 6 trit (~9.5 bits)[b] up to 6 tryte up to 3 trytes 4 trit?
1958 Electrologica X1 27 bit w 2w w w 5 bit, 6 bit
1959 IBM 1401 n c 1 c, … 1 c, 2 c, 4 c, 5 c, 7 c, 8 c c 6 bit + wm
1959
(TBD)
IBM 1620 n d 2 d, …
(4 d, … 102 d)
12 d d 2 d
1960 LARC 12 d w, 2w w, 2w w w 2 d
1960 CDC 1604 48 bit w w 12w w 6 bit
1960 IBM 1410 n c 1 c, … 1 c, 2 c, 6 c, 7 c, 11 c, 12 c c 6 bit + wm
1960 IBM 7070 10 d[c] w, 1-9 d w w w, d 2 d
1960 PDP-1 18 bit w w w 6 bit
1960 Elliott 803 39 bit
1961 IBM 7030
(Stretch)
64 bit 1 bit, … 64 bit,
1 d, … 16 d
w 12w, w bit (integer),
12w (branch),
w (float)
1 bit, … 8 bit
1961 IBM 7080 n c 0 c, … 255 c 5 c c 6 bit
1962 GE-6xx 36 bit w, 2 w w, 2 w, 80 bit w w 6 bit, 9 bit
1962 UNIVAC III 25 bit w, 2w, 3w, 4w, 6 d, 12 d w w 6 bit
1962 Autonetics D-17B
Minuteman I Guidance Computer
27 bit 11 bit, 24 bit 24 bit w
1962 UNIVAC 1107 36 bit 16w, 13w, 12w, w w w w 6 bit
1962 IBM 7010 n c 1 c, … 1 c, 2 c, 6 c, 7 c, 11 c, 12 c c 6 b + wm
1962 IBM 7094 36 bit w w, 2w w w 6 bit
1962 SDS 9 Series 24 bit w 2w w w
1963
(1966)
Apollo Guidance Computer 15 bit w w, 2w w
1963 Saturn Launch Vehicle Digital Computer 26 bit w 13 bit w
1964/1966 PDP-6/PDP-10 36 bit w w, 2 w w w 6 bit
7 bit (typical)
9 bit
1964 Titan 48 bit w w w w w
1964 CDC 6600 60 bit w w 14w, 12w w 6 bit
1964 Autonetics D-37C
Minuteman II Guidance Computer
27 bit 11 bit, 24 bit 24 bit w 4 bit, 5 bit
1965 Gemini Guidance Computer 39 bit 26 bit 13 bit 13 bit, 26 —bit
1965 IBM 1130 16 bit w, 2w 2w, 3w w, 2w w 8 bit
1965 IBM System/360 32 bit 12w, w,
1 d, … 16 d
w, 2w 12w, w, 112w 8 bit 8 bit
1965 UNIVAC 1108 36 bit 16w, 14w, 13w, 12w, w, 2w w, 2w w w 6 bit, 9 bit
1965 PDP-8 12 bit w w w 8 bit
1965 Electrologica X8 27 bit w 2w w w 6 bit, 7 bit
1966 SDS Sigma 7 32 bit 12w, w w, 2w w 8 bit 8 bit
1969 Four-Phase Systems AL1 8 bit w ? ? ?
1970 MP944 20 bit w ? ? ?
1970 PDP-11 16 bit w 2w, 4w w, 2w, 3w 8 bit 8 bit
1971 CDC STAR-100 64 bit 12w, w 12w, w 12w, w bit 8 bit
1971 TMS1802NC 4 bit w ? ?
1971 Intel 4004 4 bit w, d 2w, 4w w
1972 Intel 8008 8 bit w, 2 d w, 2w, 3w w 8 bit
1972 Calcomp 900 9 bit w w, 2w w 8 bit
1974 Intel 8080 8 bit w, 2w, 2 d w, 2w, 3w w 8 bit
1975 ILLIAC IV 64 bit w w, 12w w w
1975 Motorola 6800 8 bit w, 2 d w, 2w, 3w w 8 bit
1975 MOS Tech. 6501
MOS Tech. 6502
8 bit w, 2 d w, 2w, 3w w 8 bit
1976 Cray-1 64 bit 24 bit, w w 14w, 12w w 8 bit
1976 Zilog Z80 8 bit w, 2w, 2 d w, 2w, 3w, 4w, 5w w 8 bit
1978
(1980)
16-bit x86 (Intel 8086)
(w/floating point: Intel 8087)
16 bit 12w, w, 2 d
(2w, 4w, 5w, 17 d)
12w, w, … 7w 8 bit 8 bit
1978 VAX 32 bit 14w, 12w, w, 1 d, … 31 d, 1 bit, … 32 bit w, 2w 14w, … 1414w 8 bit 8 bit
1979
(1984)
Motorola 68000 series
(w/floating point)
32 bit 14w, 12w, w, 2 d
(w, 2w, 212w)
12w, w, … 712w 8 bit 8 bit
1985 IA-32 (Intel 80386) (w/floating point) 32 bit 14w, 12w, w
(w, 2w, 80 bit)
8 bit, … 120 bit
14w … 334w
8 bit 8 bit
1985 ARMv1 32 bit 14w, w w 8 bit 8 bit
1985 MIPS I 32 bit 14w, 12w, w w, 2w w 8 bit 8 bit
1991 Cray C90 64 bit 32 bit, w w 14w, 12w, 48 bit w 8 bit
1992 Alpha 64 bit 8 bit, 14w, 12w, w 12w, w 12w 8 bit 8 bit
1992 PowerPC 32 bit 14w, 12w, w w, 2w w 8 bit 8 bit
1996 ARMv4
(w/Thumb)
32 bit 14w, 12w, w w
(12w, w)
8 bit 8 bit
2000 IBM z/Architecture
(w/vector facility)
64 bit 14w, 12w, w
1 d, … 31 d
12w, w, 2w 14w, 12w, 34w 8 bit 8 bit, UTF-16, UTF-32
2001 IA-64 64 bit 8 bit, 14w, 12w, w 12w, w 41 bit (in 128-bit bundles)[7] 8 bit 8 bit
2001 ARMv6
(w/VFP)
32 bit 8 bit, 12w, w
(w, 2w)
12w, w 8 bit 8 bit
2003 x86-64 64 bit 8 bit, 14w, 12w, w 12w, w, 80 bit 8 bit, … 120 bit 8 bit 8 bit
2013 ARMv8-A and ARMv9-A 64 bit 8 bit, 14w, 12w, w 12w, w 12w 8 bit 8 bit
Year Computer
architecture
Word size w Integer
sizes
Floating­point
sizes
Instruction
sizes
Unit of address
resolution
Char size
key: bit: bits, d: decimal digits, w: word size of architecture, n: variable size

[8][9]

See also[edit]

  • Integer (computer science)

Notes[edit]

  1. ^ Many early computers were decimal, and a few were ternary
  2. ^ The bit equivalent is computed by taking the amount of information entropy provided by the trit, which is log _{2}(3). This gives an equivalent of about 9.51 bits for 6 trits.
  3. ^ Three-state sign

References[edit]

  1. ^ a b Beebe, Nelson H. F. (2017-08-22). «Chapter I. Integer arithmetic». The Mathematical-Function Computation Handbook — Programming Using the MathCW Portable Software Library (1 ed.). Salt Lake City, UT, USA: Springer International Publishing AG. p. 970. doi:10.1007/978-3-319-64110-2. ISBN 978-3-319-64109-6. LCCN 2017947446. S2CID 30244721.
  2. ^ Dreyfus, Phillippe (1958-05-08) [1958-05-06]. Written at Los Angeles, California, USA. System design of the Gamma 60 (PDF). Western Joint Computer Conference: Contrasts in Computers. ACM, New York, NY, USA. pp. 130–133. IRE-ACM-AIEE ’58 (Western). Archived (PDF) from the original on 2017-04-03. Retrieved 2017-04-03. […] Internal data code is used: Quantitative (numerical) data are coded in a 4-bit decimal code; qualitative (alpha-numerical) data are coded in a 6-bit alphanumerical code. The internal instruction code means that the instructions are coded in straight binary code.
    As to the internal information length, the information quantum is called a «catena,» and it is composed of 24 bits representing either 6 decimal digits, or 4 alphanumerical characters. This quantum must contain a multiple of 4 and 6 bits to represent a whole number of decimal or alphanumeric characters. Twenty-four bits was found to be a good compromise between the minimum 12 bits, which would lead to a too-low transfer flow from a parallel readout core memory, and 36 bits or more, which was judged as too large an information quantum. The catena is to be considered as the equivalent of a character in variable word length machines, but it cannot be called so, as it may contain several characters. It is transferred in series to and from the main memory.
    Not wanting to call a «quantum» a word, or a set of characters a letter, (a word is a word, and a quantum is something else), a new word was made, and it was called a «catena.» It is an English word and exists in Webster’s although it does not in French. Webster’s definition of the word catena is, «a connected series;» therefore, a 24-bit information item. The word catena will be used hereafter.
    The internal code, therefore, has been defined. Now what are the external data codes? These depend primarily upon the information handling device involved. The Gamma 60 [fr] is designed to handle information relevant to any binary coded structure. Thus an 80-column punched card is considered as a 960-bit information item; 12 rows multiplied by 80 columns equals 960 possible punches; is stored as an exact image in 960 magnetic cores of the main memory with 2 card columns occupying one catena. […]
  3. ^ Blaauw, Gerrit Anne; Brooks, Jr., Frederick Phillips; Buchholz, Werner (1962). «4: Natural Data Units» (PDF). In Buchholz, Werner (ed.). Planning a Computer System – Project Stretch. McGraw-Hill Book Company, Inc. / The Maple Press Company, York, PA. pp. 39–40. LCCN 61-10466. Archived (PDF) from the original on 2017-04-03. Retrieved 2017-04-03. […] Terms used here to describe the structure imposed by the machine design, in addition to bit, are listed below.
    Byte denotes a group of bits used to encode a character, or the number of bits transmitted in parallel to and from input-output units. A term other than character is used here because a given character may be represented in different applications by more than one code, and different codes may use different numbers of bits (i.e., different byte sizes). In input-output transmission the grouping of bits may be completely arbitrary and have no relation to actual characters. (The term is coined from bite, but respelled to avoid accidental mutation to bit.)
    A word consists of the number of data bits transmitted in parallel from or to memory in one memory cycle. Word size is thus defined as a structural property of the memory. (The term catena was coined for this purpose by the designers of the Bull GAMMA 60 [fr] computer.)
    Block refers to the number of words transmitted to or from an input-output unit in response to a single input-output instruction. Block size is a structural property of an input-output unit; it may have been fixed by the design or left to be varied by the program. […]
  4. ^ «Format» (PDF). Reference Manual 7030 Data Processing System (PDF). IBM. August 1961. pp. 50–57. Retrieved 2021-12-15.
  5. ^ Clippinger, Richard F. [in German] (1948-09-29). «A Logical Coding System Applied to the ENIAC (Electronic Numerical Integrator and Computer)». Aberdeen Proving Ground, Maryland, US: Ballistic Research Laboratories. Report No. 673; Project No. TB3-0007 of the Research and Development Division, Ordnance Department. Retrieved 2017-04-05.{{cite web}}: CS1 maint: url-status (link)
  6. ^ Clippinger, Richard F. [in German] (1948-09-29). «A Logical Coding System Applied to the ENIAC». Aberdeen Proving Ground, Maryland, US: Ballistic Research Laboratories. Section VIII: Modified ENIAC. Retrieved 2017-04-05.{{cite web}}: CS1 maint: url-status (link)
  7. ^ «4. Instruction Formats» (PDF). Intel Itanium Architecture Software Developer’s Manual. Vol. 3: Intel Itanium Instruction Set Reference. p. 3:293. Retrieved 2022-04-25. Three instructions are grouped together into 128-bit sized and aligned containers called bundles. Each bundle contains three 41-bit instruction slots and a 5-bit template field.
  8. ^ Blaauw, Gerrit Anne; Brooks, Jr., Frederick Phillips (1997). Computer Architecture: Concepts and Evolution (1 ed.). Addison-Wesley. ISBN 0-201-10557-8. (1213 pages) (NB. This is a single-volume edition. This work was also available in a two-volume version.)
  9. ^ Ralston, Anthony; Reilly, Edwin D. (1993). Encyclopedia of Computer Science (3rd ed.). Van Nostrand Reinhold. ISBN 0-442-27679-6.

On x86/x64 processors, a byte is 8 bits, and there are 256 possible binary states in 8 bits, 0 thru 255. This is how the OS translates your keyboard key strokes into letters on the screen. When you press the ‘A‘ key, the keyboard sends a binary signal equal to the number 97 to the computer, and the computer prints a lowercase ‘a‘ on the screen. You can confirm this in any Windows text editing software by holding an ALT key, typing 97 on the NUMPAD, then releasing the ALT key. If you replace ’97’ with any number from 0 to 255, you will see the character associated with that number on the system’s character code page printed on the screen.

If a character is 8 bits, or 1 byte, then a WORD must be at least 2 characters, so 16 bits or 2 bytes. Traditionally, you might think of a word as a varying number of characters, but in a computer, everything that is calculable is based on static rules. Besides, a computer doesn’t know what letters and symbols are, it only knows how to count numbers. So, in computer language, if a WORD is equal to 2 characters, then a double-word, or DWORD, is 2 WORDs, which is the same as 4 characters or bytes, which is equal to 32 bits. Furthermore, a quad-word, or QWORD, is 2 DWORDs, same as 4 WORDs, 8 characters, or 64 bits.

Note that these terms are limited in function to the Windows API for developers, but may appear in other circumstances (eg. the Linux dd command uses numerical suffixes to compound byte and block sizes, where c is 1 byte and w is bytes).

In Computers; Bits, Bytes & Words are the core concepts.

Bits – binary digits

Computer memory stores the data in bits. Each bit stores the value either 0 or 1. Inside the Computer, there are electronic switches that store the bit values; either 0 or 1. If the switch is OFF means, that represents 0 & the switch is ON means, that represents 1. That means Computer can understand the data in “0”s and “1”s. We normally use bits to represent processor architecture or design; a 32-bit processor, 64-bit processor, etc. 64-bit processor can do the computations faster than a 32-bit processor.

There must be a data transfer between Computer Memory (RAM – Random Access Memory) and the Computer Processor (CPU). The data transfer happens through BUSes. Each BUS carries an N-number of bits; depending on the Computer Architecture. Each BUS has wires to transfer the bits; 16-bit BUS carries 16-bits at a time, and it has 16 wires to transfer each bit. BUSes are used to carry the information between one Component to another Component in the Computer.

Mostly, we don’t directly deal with the Computer bits; there are computer languages that allow us to directly deal with the computer bits. “C” language provides some features to allow the programmers to directly deal with the bits.

How do we represents the data other than 0 & 1; example, characters (‘A’, ‘h’ etc,.).? The answer is Byte; another term, we use in the computer terminology is Bytes.

Bytes

A byte is a grouping of consecutive bits. Usually, 8-bits represents a Byte. Generally, we use Byte(s) to represent characters; each byte can store the values from 0-255. That means, we can store one of the characters from 0-255; in 1 Byte.

The most common term we use in Computers terminology is Bytes. Computer Programs read & write the Bytes from the Memory; also read and write the Bytes to the File Storage, and for the data processing use Bytes; file sizes will display in Bytes and etc.,

How internally Bytes store that data.? Simple, it is Bits. The Byte representation of a character ‘A‘ is:

ASCII Character value of ‘A’ is 65. Bit representation (binary representation) of ‘A’ in a Byte is below:

0 1 0 0 0 0 0 1

Note that, it is not compulsory that 8-bits are used for a Byte in all the Systems. The number of bits used for a Byte is depended on the System Architecture. Commonly we use 8-bits for a Byte.

WORDs

WORDs are also consecutive bits or bytes. This term is mostly used for CPU registers. Generally, each WORD has a length of 16-bits. CPU Registers are used to store a small piece of information while doing the calculations or processing the data. These are helpful to improve the performance of the system while doing the calculation or processing.

Note that, there are Systems that use 8-bit, 32-bit, 64-bits, 128-bits for WORDs. Commonly we use 16-bits for a WORD. The number of bits used for a WORD is purely depending on the System Architecture or design.

// Malin

16 bits.
Data structures containing such different sized words refer to them as WORD (16 bits/2 bytes), DWORD (32 bits/4 bytes) and QWORD (64 bits/8 bytes) respectively.

Contents

  • 1 Is a word 16 or 32 bits?
  • 2 Is a word always 32 bits?
  • 3 Can a word be 8 bits?
  • 4 How many bits are in a name?
  • 5 What is 64-bit system?
  • 6 What is an 8 bit word?
  • 7 What is 64-bit word size?
  • 8 How big is a word in 64-bit?
  • 9 What is word in C?
  • 10 What is a Dataword?
  • 11 How many values can 10 bits represent?
  • 12 How do you write kilobytes?
  • 13 What is a bit in a horse’s mouth?
  • 14 How many bits is a letter?
  • 15 What is a collection of 8 bits called?
  • 16 Is my computer 32 or 64?
  • 17 What is meant by 32-bit?
  • 18 Do I need 32 or 64-bit?
  • 19 What is better 8-bit or 16 bit?
  • 20 What does 8-bit 16 bit 32 bit microprocessor mean?

Is a word 16 or 32 bits?

In x86 assembly language WORD , DOUBLEWORD ( DWORD ) and QUADWORD ( QWORD ) are used for 2, 4 and 8 byte sizes, regardless of the machine word size. A word is typically the “native” data size of the CPU. That is, on a 16-bit CPU, a word is 16 bits, on a 32-bit CPU, it’s 32 and so on.

Is a word always 32 bits?

The Intel/AMD instruction set concept of a “Word”, “Doubleword”, etc. In Intel docs, a “Word” (Win32 WORD ) is 16 bits. A “Doubleword” (Win32 DWORD ) is 32 bits. A “Quadword” (Win32 QWORD ) is 64 bits.

Can a word be 8 bits?

An 8-bit word greatly restricts the range of numbers that can be accommodated. But this is usually overcome by using larger words. With 8 bits, the maximum number of values is 256 or 0 through 255.

How many bits are in a name?

Find the 8-bit binary code sequence for each letter of your name, writing it down with a small space between each set of 8 bits. For example, if your name starts with the letter A, your first letter would be 01000001.

What is 64-bit system?

An operating system that is designed to work in a computer that processes 64 bits at a time.A 64-bit operating system will not work in a 32-bit computer, but a 32-bit operating system will run in a 64-bit computer. See 64-bit computing.

What is an 8 bit word?

A byte is eight bits, a word is 2 bytes (16 bits), a doubleword is 4 bytes (32 bits), and a quadword is 8 bytes (64 bits).

What is 64-bit word size?

Bits and Bytes
Each set of 8 bits is called a byte. Two bytes together as in a 16 bit machine make up a word , 32 bit machines are 4 bytes which is a double word and 64 bit machines are 8 bytes which is a quad word.

How big is a word in 64-bit?

8 bytes
Data structures containing such different sized words refer to them as WORD (16 bits/2 bytes), DWORD (32 bits/4 bytes) and QWORD (64 bits/8 bytes) respectively.

What is word in C?

A word is an integer number of bytes for example, one, two, four, or eight. When someone talks about the “n-bits” of a machine, they are generally talking about the machine’s word size. For example, when people say the Pentium is a 32-bit chip, they are referring to its word size, which is 32 bits, or four bytes.

What is a Dataword?

A data word is a standard unit of data. For example, most IBM compatible computers have an eight bit word known as byte. Data, Measurements, Word.

How many values can 10 bits represent?

1024
Binary number representation

Length of bit string (b) Number of possible values (N)
8 256
9 512
10 1024

How do you write kilobytes?

In the International System of Units (SI) the prefix kilo means 1000 (103); therefore, one kilobyte is 1000 bytes. The unit symbol is kB.

What is a bit in a horse’s mouth?

By definition, a bit is a piece of metal or synthetic material that fits in a horse’s mouth and aids in the communication between the horse and rider.Most horses are worked in a bridle with a bit; however, horse owners who don’t care for bits will use a hackamore, or “bitless” bridle.

How many bits is a letter?

eight bits
Computer manufacturers agreed to use one code called the ASCII (American Standard Code for Information Interchange). ASCII is an 8-bit code. That is, it uses eight bits to represent a letter or a punctuation mark. Eight bits are called a byte.

What is a collection of 8 bits called?

A collection of eight bits is called Byte.

Is my computer 32 or 64?

Click Start, type system in the search box, and then click System Information in the Programs list. When System Summary is selected in the navigation pane, the operating system is displayed as follows: For a 64-bit version operating system: X64-based PC appears for the System Type under Item.

What is meant by 32-bit?

32-bit, in computer systems, refers to the number of bits that can be transmitted or processed in parallel.For microprocessors, it indicates the width of the registers and it can process any data and use memory addresses that are represented in 32-bits.

Do I need 32 or 64-bit?

For most people, 64-bit Windows is today’s standard and you should use it to take advantage of security features, better performance, and increased RAM capability. The only reasons you’d want to stick with 32-bit Windows are: Your computer has a 32-bit processor.

What is better 8-bit or 16 bit?

In terms of color, an 8-bit image can hold 16,000,000 colors, whereas a 16-bit image can hold 28,000,000,000. Note that you can’t just open an 8-bit image in Photoshop and convert it to 16-bit.More bits means bigger file sizes, making images more costly to process and store.

What does 8-bit 16 bit 32 bit microprocessor mean?

The main difference between 8 bit and 16 bit microcontrollers is the width of the data pipe. As you may have already deduced, an 8 bit microcontroller has an 8 bit data pipe while a 16 bit microcontroller has a 16 bit data pipe.A 16 bit number gives you a lot more precision than 8 bit numbers.

Skip to content

Difference between Bit, Byte and Words

This article will help you to learn about the difference between bit, byte and words.

Difference between Bit, Byte and Words

Bit

The computer does not have a large number of symbols for representing data. It has only two, 0 and 1 (called binary digits or bits). These correspond to the two electronic ox magnetic states used in computer circuits and storage.

The smallest unit used for feeding data and program into computer is bit. Information is handled in the computer by electrical components such as transistors, integrated circuits, semiconductors and wires, all of which can only indicate two states or conditions.

Transistors are either conducting or non-conducting; magnetic materials are either magnetized or non-magnetized in one direction or in the opposite direction; a pulse or voltage is present or not present.

These two possible states can be expressed with the help of bits -0 and 1. For example, the presence of current pulse in a circuit in computer may be represented by the bit 1 and the absence of current pulse in a circuit may be represented by the bit 0.

Byte

A collection of some bits is called a byte. Byte is a group of adjacent bits (binary digits) operated upon as a unit. An 8 bit unit is commonly called a byte and has become the standard unit for storing a single character.

In many computers, it is 8-bit set encoding one alphanumeric character or two decimal digits. Alphanumeric is a contraction for alphabetic (A, B, C, etc.) and numeric, (0, 1, 2, etc.). A set of alphanumeric characters will usually include special characters too such as dollar sign, comma etc.

Words

Some memory units are not made up of bytes but of words. A computer word consists of the data which is stored or retrieved when a memory location is specified Word is a collection of bits treated as a single unit. Word is an ordered set of characters handled as a unit by the computer. The word may be fixed or variable in length. The word length depends upon the number of bits or characters in a word.

The number of bits varies from 4, 8, 12, 16, 32 etc., up to 64 i.e., the word may be as long as 64 bits or as short as 8 bits. In a fixed word-length computer, the number of characters in a word does not vary, and an address will typically refer to one set of characters. In a variable word-length computer, each character or byte has an address and the word utilized by the computer can include a variable number of characters.

The length of the variable word is specified either by the instruction which calls for it or by a word- mark in storage. A byte is usually shorter than a word, typically consisting of 8 bits. In some computers the grouping of bits, bytes or words is flexible in design to meet the different storage requirements of numbers, alphanumeric characters and instructions.

Page load link

Go to Top

BITS, BYTES AND WORDS

BITS: 

The bit is the smallest value which can be stored in a computer. The bit is also called a binary digit. The bit can hold binary value either 0 or 1.

Nibble: 

The nibble is a collection Of four bits.

Byte: 

The byte is a collection Of eight bits.
The byte is the standard unit Of measurement Of computer memory data storage and transmission speed.

A DVD-ROM has a capacity of 4.7 gigabytes, and internet speeds Of 4 megabits per sec(MBPS) is common.

Length Name Example
1 bit 0
4 nibble 1011
8 bytes 00101100

The relationships between these units are expressed in the following manner:

relationships between these units

Words: 

Even though computer memory is measured in bytes (rather mega- or gigabytes), the CPU handles memory data in larger units, called words, where a word is usually an even multiple Of bytes (two bytes, four bytes etc.). This number purely based on processor architecture.




Most machine instructions are one word in size, and because a CPU register must be able to hold an instruction, it too is one word wide.


Video tutorial:

Definitions

Bit = Binary digIT = 0 or 1

Byte = a sequence of 8 bits = 00000000, 00000001, …, or
11111111

Word = a sequence of N bits where N = 16, 32, 64 depending on the
computer

There are 2N words of length N. Why?

How many bytes are there?

Measuring Amount of Data/Memory Capacity

1 kilobyte = 1 KB = 210 bytes = 1024 bytes (appx 103
bytes)

1 Megabyte = 1 MB = 220 bytes = 1,048,576 bytes (appx
106 bytes)

1 Gigabyte = 1 GB = 230 bytes = 1,073,741,824 bytes
(appx 109 bytes)

What are some other prefixes?

Bandwidth, the
speed at which data can be transmitted over a given communication channel, is
measured in bits/second.

Representing Bits

A bit can be represented as a voltage in a hardware
computer. For example, 3 � 5 volts = 1, 0 � 2 volts = 0.

We can represent Boolean truth values in a virtual computer
using bits: 1 = true, 0 = false.

A single Binary Digit is known as a bit, and can
represent either the value 0 or
1. Bits can be implemented in
computer hardware using switches. A simple way to think about it is,
that If the switch is on then the bit is 1 and if the switch is off
then the bit is 0.  Of course, the underlying electrical
engineering is much more involved than this, but this is the essential
idea.

A byte is a sequence of bits. Since the mid 1960’s
a byte has been standarized to be 8 bits in length. 01000001
is an example value that may be represented by a single byte.
Since there are 8 bits in a byte there are 28 different
possible sequences for one byte, ranging from 00000000
to 11111111. This means that a byte can be used to
represent 28, or  256 distinct values.  Like
bits, bytes too may be used in sequence to allow for more
possibilities.  Two contiguous bytes allows for  216,
or  65,536 distinct values to be represented, while 4 contiguous
bytes allows for 232, or  4,294,967,296 distinct values.

Computer keyboards have a limited number of keys and even with the
multiple values provided by the use of the Shift, Alt, Ctrl and other such keys,
the number of distinct key values is less than 256.  Thus,
every keystroke may be represented by a unique one byte binary
value. 

Since each character (letters, decimal digits and special
characters such as punctuation marks, etc) can be represented by a
single byte value, a standard is needed to insure that the mapping of
characters to byte values is consistent across computer systems.
 There are two standard codes that use one byte to represent a
character, ASCII
(ass’-key) and EBCDIC
(ib’-suh-dik). ASCII, the American Standard Code for
Information Interchange, is the code that is most commonly used today.
EBCDIC, Extended Binary Coded Decimal Interchange Code, was used by IBM
on its large mainframe computers in the past. Since these codes are
limited to 256 possible combinations, certain character sets, such as
Chinese, Arabic, Japanese, Klingon and others, cannot be represented
using these codes. This problem has been solved by developing
another standard, Unicode,
which uses 2 bytes for each character. This extension allows 216
different symbols to be represented, a total of 65,536. The use of
Unicode provides for international standardization and uniformity but
consumes twice the amount of computer resources.  Each character
requires two bytes of memory to be represented and likewise twice the
number of bytes need to be transmitted across communication channels.

The term word refers to the
standard number of bits that are manipulated as a unit by any
particular CPU. For decades most CPUs had a word size of 32 bits
(or 4 contiguous bytes), but word sizes of 64 bits are becoming
more and more commonplace. The signifcance of the word size of a
particular computer system is that it reflects the amount of data that
can be transmitted between memory and the processor in one chunk.
 Likewise, it may reflect the size of data that can be manipulated
by the CPU’s ALU in one cycle.  Computers can process data of
larger sizes, but the word size reflects the size of the data values
the computer has been designed to readily process directly.  All
other things being equal, (and they never are), larger word size
implies faster and more capable processing.

Further Information

Wikipedia contains exhaustive information about bits, bytes and
words.

howSTUFFworks also has an article on (How
Bits and Bytes Work)

Понравилась статья? Поделить с друзьями:
  • Word by word английские слова по карточкам
  • Word by word speed reading
  • Word by word workbook скачать
  • Word by word she spoke to me
  • Word by word second edition pdf