# Useful Notes: Powers of Two Minus One

Video games are nothing more than computer programs, and they're limited by the systems they run on. One particularly strict limitation is the fact that computers can't handle numbers of arbitrary size with any kind of efficiency. Because efficiency is normally pretty important for video games, this places some hard limits on the range of values they can work with. Because computers work in binary, that size has to be a certain number of bits (binary digits — a zero or a one). Numbers stored in a computer's memory are called

*variables*. Computers also have a limit to the maximum size a variable can be. It's not a hard limit, but using bigger numbers slows down computation considerably. Whenever you hear about the "number of bits" of a video game console (e.g., the Nintendo Entertainment System being "8-bit"), that's what they're referring to. The NES has an 8-bit data bus; it can only move 8 bits of data around at a time. If it has to move any more, it has to do it in more than one cycle. In technical terms, this is called "word size." In newer video games, many things — say, the number of items in the player's Hyperspace Arsenal — are capped at reasonable limits because the variable actually used to store that value has a decidedly*un*reasonable limit. However, in older days, it was common to let the computer itself cap these values, because the limits were a lot lower — an 8-bit computer such as the NES has a "limit" of 255. The highest value you can store in a variable or register can be determined by the number of bits it uses. It's 2 to the power of the number of bits, minus 1 because we're counting from zero. Here's a handy table.-
**8 bits**can hold a value as high as 255. The Atari 2600 and 7800, Nintendo Entertainment System, Sega Master System, and early home computers were all 8-bit. -
**12 bits**can hold a value as high as 4095. Some NES games like*Final Fantasy I*used hacks to get values this large. -
**16 bits**can hold a value as high as 65,535. The Super Nintendo Entertainment System and Sega Genesis are 16-bit, as are PCs made in the mid-1980s, and a few Japanese gaming computers. -
**32 bits**can hold a value as high as 4,294,967,295. The Sega Saturn and Dreamcast, Sony PlayStation, Microsoft XBox and the Wii are all 32-bit. So are desktop PCs made from 1987 to 2007. Most smartphones are 32-bit, with 64-bit phones starting to trickle out in 2013 and 2014. -
**64 bits**can hold a value as high as 18.4 quintillion (or 1.84×10^{19}). The Nintendo 64, Game Cube^{note }, Sony PlayStation 2, and modern desktop PCs are all 64-bit, with 64-bit operating systems becoming rather commonplace in 2010 (at least, on new computers). -
**128 bits**can hold a value as high as 340.2 undecillion (or 3.4×10^{38}). Computers usually do not have 128-bit general-purpose registers, but the PS3's Cell, the X-Box 360's CPU, and modern x64 CPUs (used both in PCs and Macs) do have 128-bit vector registers (holding 4 32-bit values). These are used for vector math.

*and*negative numbers. Zero is taken to be the first "positive" number, whereas negative numbers start at -1, which makes the negative limit higher (absolutely speaking). This effectively reduces by half the maximum number that the value can hold for the sake of representing negative numbers. Here are some revised tables:-
**8 bits**can hold a value from -128 to 127. -
**12 bits**can hold a value from -2048 to 2047. -
**16 bits**can hold a value from -32,768 to 32,767. -
**32 bits**can hold a value from -2,147,483,648 to 2,147,483,647.

*larger*than the hardware word size, potentially with millions of digits, but this is rarely needed in video games (and is*significantly*slower, too). So that explains the numbers themselves. But what causes them to crop up in games? There are a few reasons these numbers show up:- As a Cap. Programming for older systems is really
*really*hard; these systems were relatively good at graphics processing, but very bad at calculations, so every CPU cycle saved was a big deal.^{note }If you cap a value to the actual limit of your variable, you can enforce the cap with a "did previous instruction overflow" test, which is usually faster than a "is X greater than Y" test (the latter often involves hidden subtraction, even on newer processors).- Note that many games exhibit bugs due to
*not*checking the overflow after addition. That is, there's a cap because once you cross the cap, something breaks or behaves strangely.

- Note that many games exhibit bugs due to
- To use shifts. Computers can multiply and divide by powers of two by using left- and right-shift operations. Shifts are much,
*much*faster than multiplication or division, but only work for powers of two.- Programmers in the early days found ways around that. Want to multiply by 9? Shift left three times and add on the original number. This trick is expanded upon and used by some modern processors.
- It might backfire on modern processors as while multiplication takes a few cycles but the shifts would introduce data hazards and cause stall cycles. Hopefully the processor will implement out-of-order execution.

- Programmers in the early days found ways around that. Want to multiply by 9? Shift left three times and add on the original number. This trick is expanded upon and used by some modern processors.
- Errors. Many function calls use -1 as an error code. If this value is not checked for, and then gets stored into an unsigned variable, it becomes the highest number the variable can hold (due to how "two's complement" signing works.)
- Divide By Zero. As a specific example of the above, some older systems would return the highest number the variable can hold as the result of division by zero (think of it as "infinity"). This is much more desirable than crashing when the player's save file is on the line.
- As of Divide By Zero for floating point numbers the behaviour is mandated by IEEE 754 (a standard of handling floating point numbers). Sometimes this makes things worse if it hides an error - instead of clear crash during test phase you have a potentially corrupted save file.

- For alignment purposes. For most purposes, data in memory has to be aligned to its size. For example, a one-byte (8-bit) value can be stored at any address, but a four-byte value must be stored on some platforms at an address that is a multiple of four. When game programmers used to map out memory used by the game, they would keep in mind this alignment requirement. Say a given block of memory is used to store player inventory, and each item in the inventory is a single byte. It makes sense to give the player a power of two items, since any value stored after it will have the same alignment as the start of the block.
- Reading in data for something it's not meant for. A very frequent source of Good Bad Bugs in olden times was when data was loaded as something that it actually wasn't. For example, suppose something goes iffy in the code, and the game starts reading map data as if it were attack data. But map data is something very different from attack data. Perhaps the game uses the value FFFF (65535 in hexadecimal) to indicate the end of the map. This is a common value to use for this purpose (called a "terminal value" in programming jargon) because it's the
*last*value two bytes can store, so it's easy to remember. (One FF by itself might be used for, say, the end of a room, or the end of a script that runs on that map.) If the code starts reading in a map as attack data, however, these values may end up anywhere. If FFFF were read in as the damage of an attack, for example, the attack would deal 65535 damage (possibly capped to 9999, of course). - Because programmers have these numbers etched into their mind— to a sufficiently binary-minded programmer, "256" (binary 100000000) is more of a round number than "300". Hence they show up for no reason at all.
- That's a part of the reasoning for the whole concept of "magic numbers" in engineering. When you are making something you need to decide on some properties (like plank size or ceiling height, for example) anyway, so why not use the ones that someone has already used before? And then it happens that the similar sizes allow for compatibility and interoperability, it becomes a self-reinforcing convention, and we all end up with 2-by-4-s. The same logic applies to the programming too.

- As a naming convention. Everything in a computer indexes, or starts, from position 0. For example, in an 8-bit number, bit 0 is the first bit. How does this make sense? Because if it's a 1, the value of the number is 2^0, or 1.

- Fixed point. This means you're essentially taking your number and sticking the decimal point at a specific spot in that number. The point never moves around. On the plus side, these are very fast to work with. Note that the number is still in binary, so it's about halves and quarters and such, rather than the tenths and hundredths and such that decimals use.
- Floating point. This works like scientific notation; one bit is used to store the sign, some number of bits (called the mantissa) are used to store the number to the right of the decimal point, and the remaining bits (called the exponent) are, predictably, used to store the power of 2 to multiply by. (We use 2 instead of 10 because, again, computers work in binary.) It should be noted that this is a gross oversimplification.