Boolean Logic and Digital Circuits
Modern digital computers are built from digital logic circuits whose basic building blocks are logic gates, each of which is designed to implement a specific logical function. Information is held in data "words", representing data or instructions, made up from strings of individual "bits" with binary values of 1 or 0. These values are analogous to Boolean logic propositions and the statements or conclusions derived from them which were defined as "true" or "false". Boolean algebra is the tool used to design combinations of gates to implement more complex functions such as mathematical operations, control functions and data storage.
Boolean Algebra
Boolean algebra is based upon a twovalued, or binary scheme. The two values may be expressed in many ways, such as true or false, 1 or 0, and "on" or "off". It is this property which was recognised and developed by Claude Shannon in 1937 which makes it so useful for implementing logic functions by means of electronic circuits. For example, logic 1 and logic 0 might be implemented as two different voltage levels in a circuit, or by the status of a switch, or by the presence or absence of a current in the circuit.
Notation
The engineering application of Boole's logic uses a simplified version of the original notation as follows,
 A logical OR is equivalent to Boolean addition and is represented by a plus + sign as in A+B ≡A OR B.
It can represent parallel switch contacts.
 A logical AND is equivalent to Boolean multiplication which is represented by a dot . sign as in A.B ≡ A AND B.
It can represent series switch contacts.
Note that it has become conventional to drop the dot . sign (AND symbol) so that A.B is written as AB.
 A logical NOT is equivalent to Boolean complementation or negation and is represented by an overbar ‾‾ above the relevant variable as in ≡ NOT A.
It can represent normally closed switch contacts.
 The Exclusive OR or logical XOR was not an original Boolean operator and has its own special symbol as in ≡ B + A which is true if A or B is true but NOT both.
Boolean Laws
Duality
Note that every law has two expressions, (a) and (b) which are duals of each other. Duality means
 Changing every OR (+) operation of the expression to an AND (.) and every AND (.) to an OR (+).
 Changing all the 0 elements to 1's and viceversa.
Commutative Law
(a) A + B = B + A
(b) A B = B A
Associate Law
(a) (A + B) + C = A + (B + C)
(b) (A B) C = A (B C)
Distributive Law
(a) A (B + C) = A B + A C
(b) A + (B C) = (A + B) (A + C)
Identity Laws
(a) A + A = A
(b) A A = A
(a) AB +A = A
(b) (A+B)(A+) = A
Redundance Laws
(a) A + A B = A
(b) A (A + B) = A
(a) 0 + A = A
(b) 0 A = 0
(a) 1 + A = 1
(b) 1 A = A
(a) A+ = 1
(b) A = 0
(a) A+B = A+B
(b) A(+B) = AB
Involution Law
(a) = A
De Morgan's Theorem
(a) = (Breaking the Overbar changes the OR to an AND)
(b) = + (Breaking the Overbar changes the AND to an OR)
Note: is not the same as
In addition to the above Boolean algebra, digital logic gates have been developed to represent Exclusive OR and Exclusive NOR expressions extending the original range of Boolean laws. See exclusive OR expressions below.
Logic Gates
Logic gates may have two or more inputs and, except in some special cases, they have a single output. The status of the input and output terminals can only be in one of the two binary conditions, either low (0) or high (1), represented by two different voltage levels, typically 0 volts for logic 0, and around 3 to 5 volts for logic 1, depending on the semiconductor technology used. Logic gates also require a power supply.
The Transistor as a Switch
Electronic gates are generally constructed from transistor circuits which depend for their operation on the use of the transistor as a switch, rather than as an amplifier for which it was originally invented. With no voltage on the base, there is no current through the transistor which is thus switched off and the output (collector) voltage will be high. When a "high" voltage is applied to the base the transistor is switched on and output (collector) voltage will be "low". See more on the page about semiconductors.
An early version of a bistable switching circuit was the 1919 Eccles and Jordan flipflop based on valves (vacuum tubes). The later transistor version was one of the first electronic circuits to be implemented as an Integrated Circuit by Robert Noyce in 1959.
Flipflops rely on the concept of feedback in which the output of a circuit is fed back into the input such that when the input is high, the output is low and vice versa.
See below an example of transistor switches used in the electronic circuit used to implement a three input NOR gate 

Logic Gates and Truth Tables
Logic circuit truth tables show the status of the output terminal or terminals of logic gates and logic circuits for all the possible input combinations. The gate or circuit's input states are shown in the left columns of the table while the corresponding output states are shown in the right columns.
The tables opposite show the range of common logic gates with their corresponding truth tables. 
Logic Gates 
Truth Tables 
AND Gates

A

B

A.B


0 
0 
0 
1 
0 
1 
0 
1 
1 
0 
0 
1 
1 
1 
1 
0 

OR Gates

A

B

A+B


0 
0 
0 
1 
0 
1 
1 
0 
1 
0 
1 
0 
1 
1 
1 
0 

Exclusive OR Gates

A

B



0 
0 
0 
1 
0 
1 
1 
0 
1 
0 
1 
0 
1 
1 
0 
1 


A


0 
1 
1 
0 


XOR and XNOR Gates
The Exclusive OR (XOR) gate with inputs A and B implements the logical expression = B + A
 When both the inputs are different, then output becomes high or logic 1.
 When both the inputs are same, then output becomes low or logic 0.
The Exclusive NOR (XNOR) gate with inputs A and B implements the logical expression = = AB +
 When both the inputs are same, then output becomes high or logic 1.
 When both the inputs are different, then output becomes low or logic 0.
Exclusive OR gates are commonly used for comparisons and parity checks.
The Vernam Cipher
The Vernam cipher is a special application of XOR logic. Also called Modulo 2 Addition, it is similar to a digital adder except that the carry digits are ignored.
An important cryptography tool, its special property is that a plaintext message string can be enciphered by XORing it with a random symbol scrambler string or key of the same length to create truly unbreakable ciphertext. The ciphertext can however be deciphered directly by XORing it with the original scrambler key.
Vernam used the five bit Baudot code to represent each character with the original notation of + and  to represent the logic states, rather than the more familiar 1 and 0 used today. In the example opposite:
 The plaintext letter A is enciphered by adding a random scrambler letter, for example letter B. This generates the Baudot code for ciphertext G.
 By adding the same scrambler letter B to the ciphertext G the original plaintext letter is regenerated.

Truth Table 
Inputs 
Output /
Input 
Output 
Plaintext 
Scrambler 
Ciphertext 
Plaintext 
A 
B 
= G 
=A

+

+ 
 
+

+ 
 
+ 
+

 
 
 


 
+ 
+ 




+

+




Three Input NOR Gate
The diagram opposite is an example of a three input NOR gate showing the electronic circuit from which the gate is constructed together with the circuit symbol for the gate and the truth table associated with the gate.
Applying a logic 1 to any of the input terminals A, B or C cause a current to flow through the load resistor R4 which in turn causes the voltage on the output terminal Z to fall to logic level 0.
Only when all the input terminals are set to logic 0 will the current through the load resistor be cut off and the voltage on the output terminal will rise to logic level 1.
Thus there can only be a logic 1 output if neither, A nor B nor C are set to logic 1. 


Z = 
Truth Table

A 
B 
C 

0 
0 
0 
1 
0 
1 
0 
0 
0 
0 
1 
0 
0 
1 
1 
0 
1 
0 
0 
0 
1 
1 
0 
0 
1 
0 
1 
0 
1 
1 
1 
0 

Digital Logic Circuits
Boolean logic is used to design complex digital circuits to perform a wide variety of logical functions. There is however often more than one way to implement a logic circuit by using alternative types of gates. Some examples follow.
SetReset FlipFlops and Latches
The SetReset flipflop constructed from two cross connected, two input, NOR gates is one of the fundamental digital logic circuits. It is a bistable circuit which can store a single data bit in the form of a binary zero or a binary one and is used as a memory device or a latch.
Applying a logic1 to the Set terminal S stores a 1 and sets the output terminal Q to logic 1.
Applying a logic 1 to the Reset terminal R clears the memory, storing a 0 instead and sets Q to logic 0.
is the inverse or complement of Q
Flipflops or latches can also be constructed from NAND gates with similar cross connections.

SetReset FlipFlop with NOR Gates 

Truth Table 
Set/Reset 
Output 
S 
R 
Q 

Result 
0 
0 


No change 
0 
1 
0 
1 
Latch Reset 
1 
0 
1 
0 
Latch Set 
1 
1 


Not allowed 


Registers are common storage devices providing temporary storage of multibit data words such as 4, 8 or 16 bit words. They are constructed from groups of flipflops each storing a single bit of information so that n flipflops are used to store an n bit word.
Adders
The circuit opposite is a single bit half adder built from five NAND gates.
Half adders have only two inputs for the two bits to be added and can not accept a carry bit from a previous stage. They do however have two outputs, one for the Sum and the other for the Carry output from the two bit addition.


Truth Table 
A 
B 
Sum 
Carry 
0 
0 
0 
0 
1 
0 
1 
0 
0 
1 
1 
0 
1 
1 
0 
1 

The half adder opposite is another example of how logic functions can be implemented in different ways. In this case the adder circuit can be simplified by using only two gates, an AND gate and an XOR gate to perform the same half adder function as the circuit above. 

Full adders are designed to accept a carry bit from a previous stage and hence have three inputs. The circuit below is a an example of single bit, full adder constructed entirely from two input NOR gates. In this case it is essentially two, twoinput, half adders in series with the input carry bit bypassing the first adder and being added to the sum of the two input bits from the first adder, in the second adder.
Note that it takes 12 such gates simply to add two single bits plus any input carry bit from a previous addition stage and to provide two output bits representing the sum of the bits and any associated carry bit. A logic circuit designed to add two eight bit words will require eight times as many gates.

Truth Table 
Inputs 
Outputs 
A 
B 
C_{in} 
Sum 
C_{out} 
0 
0 
0 
0 
0 
0 
0 
1 
1 
0 
0 
1 
0 
1 
0 
0 
1 
1 
0 
1 
1 
0 
0 
1 
0 
1 
0 
1 
0 
1 
1 
1 
0 
0 
1 
1 
1 
1 
1 
1 

It may seem strange to use so many gates when the circuit could easily be implemented with fewer, more complex gates, but circuits such as the adder above were used in the Apollo Guidance Computer which took the US astronauts to the Moon in 1969. All of its digital circuits were built from three input NOR gates. This was because they needed highly reliable semiconductor components and at the time (1966) when the computer design was frozen, integrated circuit technology was still in its infancy and NASA wanted to limit the number of different components used to those which had a proven track record. NOR gates were chosen because they were one of the very few options that met this requirement and because NOR gates were simple and more versatile than other available gates for building more complex functions.
Binary Arithmetic
A binary adder can be adapted to perform other arithmetic operations such a subtract, multiply and divide as well as other more complex mathematical functions, avoiding the need for multiple specialist processors, by making use of the following principles of binary arithmetic.
 A 1's complement of a number is the same as the number with all its 1's changed to 0's and all its 0's changed to 1's. This move enables the adder to deal with negative numbers and subtraction.
 A 2's complement is the same as a 1's complement with the addition of an extra 1 to the least significant bit.
 The sign of a positive or negative binary number is changed by taking its 2's complement.
 A left shift of all the bits in a binary number by 1 position is the same as multiplying the number by 2 (binary 10).
 A right shift of all the bits by 1 position is the same as dividing the number by 2 (binary 10).
 A number can be raised by the n th power of 2 by adding n zeros to the end of the number.
 Multiplying a multibit number by 1 results in the same number. Multiplying it by 0 results in 0 and can be ignored. This makes multiplication very simple.
 Dividing a multibit number 1 returns the same number.
 Dividing any number 0 is not permitted.
 Multiplication of two multibit numbers involves repeating the "left shift and multiply" operations n times in a loop, where n is the number of bits in the multiplier.
 Division of two multibit numbers involves repeating the "right shift and divide" operations mn times in a loop, where m is the number of bits in the dividend (the number being divided) and n is the number of bits in the divisor.
 Subroutines carry out a series of instructions to perform arithmetic operations.
The following are three of the most common examples.
 Subtraction
To subtract binary number B (the subtrahend) from binary number A (the minuend).
 Add the 2's complement of B to A
 If there is a final carry bit, discard it and the result is positive.
 If B greater than A there will be no carry bit and the result will be the 2's complement of the sum and is negative.
 (There's a slightly different way of making the subtraction with the 1's complemet of B instead of the 2's complement.)
 Multiplication
This involves n steps where n is the number of bits in the multiplier.
The first Step
 Start with the least significant bit (LSB) of the multiplier.
 If this multiplier bit is 1, add the multiplicand (the number being multiplied) to the product (the result of the multiplication). (The initial value of the product will be zero)
 If this multiplier bit is 0, ignore and move to the next step.
In each subsequent step:
 The multiplicand is shifted one place to the left.
 The next bit of the multiplier is examined.
 If this bit is 1, the shifted multiplicand is added to the current value of the product.
 If this bit is 0, ignore and move to the next step.
 Repeat the step n times in a loop (until the multiplicand has been multiplied by every bit of the multiplier).
 Division
This operation involves mn steps where m is the number of bits in the dividend (the number being divided) and n is the number of bits in the divisor.
Note: Checks must be included and the operation set up to avoid the potential problem of dividing by 0.
 Start by aligning the most significant bit (MSB) of the divisor directly below the MSB of the dividend.
 Compare the n bits of the divisor the corresponding n bits of the dividend directly above.
 If the divisor is larger than the dividend, set 0 as the first bit of the quotient (the result of the division)
 If the divisor is smaller than the dividend, subtract the divisor from the dividend to get new dividend/remainder and set the first bit of the quotient to 1.
 Move to the next step
In each subsequent step:
 Shift the divisor one place to the right and compare it with next n bits of the dividend
 If the divisor is larger than the corresponding n bits of the dividend, set the next bit of the quotient to 0
 If the divisor is smaller than the dividend, subtract the divisor from the dividend to get new dividend/remainder and set the next bit of the quotient to 1.
 Repeat the step mn times in a loop (until the LSBs of the divisor and the dividend line up)
 The result of the subtraction in the last step is the remaider of the division.
The subroutines include special check points to detect and insert extra instructions to deal with carry and borrow bits, overflows, negative numbers, remainders, divide by zero whether the divisor is greater than the dividend.
Subroutines similar to those above, in combination Boolean logic circuits, are used to enhance the capability of the computer's Arithmetic Logic Unit (ALU) to enable it to carry out many more complex functions.
See more about Computer Architecture
Floating Point Arithmetic
The computer's calculating unit has just two hardware tools in its tool bag, "shift" and "add" to carry out all its mathematical operations. The basic four function arithmetic operations, outlined in the foregoing section, were all carried out with these two tools and all concerned operations on integer numbers. However in practical applications, even the most simple mathematical calculations involve operating with decimal numbers.
Besides this, the computer may also be required to handle transcendental functions such as logarithms, exponential and trigonometric functions which can not be represented by simple algebraic expressions and can thus not be processed by the computer's simple shift and add capability. Transcendental functions can however be expanded and expressed as an infinite series of simpler algebraic expressions each of which can be processed by the machine's basic operations but summing an infinite series is not practical. Some form of mathematical approximation is required. Fortunately, the initial algebraic terms of the series usually, but not always, converge quickly down to very small values such that subsequent terms can safely be ignored without seriously affecting the accuracy of the result. Thus the transcendental function can be considered as a polynomial consisting of just the first few significant terms in the series. See examples (below)
In addition to the problem of precision, there's also the problem of scale. It is difficult to manage both very large and very small numbers in computer data registers and communications buses of practical size. Some examples:
 The mass of an electron is about 0.00000000000000000000000000000091093836 kg. In scientific notation, this is written 9.1093836 x 10^{31} kg.
 The Earth's mass is about 5972400000000000000000000 kg. In scientific notation, this is written 5.9724 x 10^{24} kg.
 The speed of light is about 300,000,000 m/s (3.0 x 10^{8} m/s)
 Newton"s gravitational constant is about 0.0000000000667 Nm^{2}kg^{2 } (6.67 X 10^{11 }Nm^{2}kg^{2 })
These last two numbers may even appear in the same equation.
The size problem is even more difficult when the numbers are represented in their much longer binary form.
The use of conventional scientific notation solves part of this problem but introduces the need to manipulate the position of the decimal point.
These problems of accuracy, precision and scale are solved by the use of floating point arithmetic which uses the more convenient scientific notation but calculation methods are still limited to the computer's basic shift and add capability. These basic operations can however be augmented by software subroutines and extra logic and registers to store temporary or intermediate values. The following are some examples of floating point calculations:
First some definitions:
Number Definitions
 Mantissa
Also called the Significand, contains all the number's digits. Negative significands represent negative numbers. (Beware that the term "mantissa" may cause confusion because it can also refer to the fractional part of the common logarithm.)
 Base
The reference number on which the exponent is based.
 Exponent
It is the power to which the base is raised.
It indicates where the decimal (or binary) point is placed relative to the beginning of the significand.
 Radix
The binary equivalent to the decimal point in decimal numbers.
Example
In scientific notation the decimal number 345.6 can be represented by 3.456 X 10^{2} where 3456 is the mantissa or significand, 10 is the base and 2 is the exponent.
Floating Point Number Benefits
 They can represent numbers of widely different magnitudes (Range is limited by the length of the exponent)
 They provide the same relative accuracy at all magnitudes (Accuracy is limited by the length of the significand)
 In calculations involving numbers with both very large and very small magnitudes they enable the accuracy of both to be preserved in the result.
Floating Point (FP) Operations
Floating point operations can be implemented by hardware (circuitry) or by software (program code). Software is however much slower than hardware by two or three orders of magnitude.
Bit Shifting
Shifting the mantissa left by 1 bit decreases the exponent by 1 and moves the radix point right by one place.
Shifting the mantissa right by 1 bit increases the exponent by 1 and moved the radix left by one place.
Basic FP Algebraic Functions
 FP Addition
 Align the decimal points by increasing or decreasing one of the exponents so that both exponents are the same. This will be the new mantissa for the sum.
 Change the corresponding mantissa accordingly.
 Add the new mantissas to get a new mantissa for the sum.
 FP Subtraction
 Align decimal points and change the mantissa as above
 Subtract using 2's complement
 FP Multiplication
 Multiply the mantissas to get the new mantissa
 Add the exponents to get the new exponent
 FP Division
 Divide the mantissas to get the new mantissa
 Subtract the exponents to get the new exponent
Transcendental and Other Functions
It is possible, but not necessarily practical, to store all the required values of transcendental functions in lookup tables stored in the computer's memory. While this would be convenient, in many applications however, it would require impractically large memories. Instead the necessary values must be calculated as required.
Estimating the value of a transcendental function involves setting up a loop to calculate the value of each significant term in the series in turn, and summing the values in a separate accumulator (taking the sign into account). The more the terms in the series, the better will be the accuracy, but the processing time will be correspondingly longer due to the increased number of calculations.
It is often possible to represent the transcendental function using alternative mathematical approximation methods. As in the tradeoff between speed and accuracy above, the alternative methods also involve similar tradeoffs. The Taylor series is a common mathematical expansion, though not the only one, used to approximate the value of transcendental functions. Some typical examples using the Taylor series follow.
See more about the Taylor series.
See also the CORDIC expansion (below).
 Trigonometric Functions
 The following series are the Taylor expansions for the sine and cosine where the variable X is measured in radians.
sin(X) 
≈ 
X 

X^{3} 
+ 
X^{5} 

X^{7} 
+ 
. . . 
+ 
(1)^{i} 
X ^{(2i+1)} 
+ 
. . . 
3! 
5! 
7! 
(2i + 1)! 
cos(X) 
≈ 
1 

X^{2} 
+ 
X^{4} 

X^{6} 
+ 
. . . 
+ 
(1)^{i} 
X^{(2i)} 
+ 
. . . 
2! 
4! 
6! 
(2i)! 
Note that these two series involve both positive an negative terms and the denominators involve increasing factorials. This means that the terms will converge very quickly to small magnitudes.
Note also that the computer does not usually calculate the numerical values of the factorials which are instead retrieved from lookup tables in the memory.
An alternative to the Taylor series for estimating the value of the sine function is the Hastings approximation which is about three times faster and only slightly less accurate.
The function is divided into a small number of intervals and, in each of these, a straight line approximation is used. The slopes of the straight line segments have denominators which are powers of 2, and so the calculation does not need floating point multiplication or division operations. This was the algorithm used for trigonometric calculations in the Apollo 11 Guidance Computer which took the astronauts to the Moon.
 Logarithms
 Exponential
 This is the Taylor expansion for the exponential (e^{z})
exp(Z) 
≈ 
1 
+ 
Z 
+ 
Z^{2} 
+ 
Z^{3} 
+ 
Z^{4} 
+ 
Z^{5} 
+ 
. . . 
+ 
Z^{i} 
+ 
. . . 
2! 
3! 
4! 
5! 
(i)! 
Note that all the terms in this series are positive and the numerators increase more quickly than the denominators so that the series does not converge but expands.
 Square Root
The square root is not a transcendental function. Numerous ways for calculating square roots have been developed since the first formal method was proposed by the Greek mathematician Hero of Alexandria in the first century A.D.
 Iterative Method
The simple method based on "guess, check and refine" has been used for many years. It works as follows where a is the desired square root √x
 Guess an answer "a" between 0 and x.
 Calculate a^{2}
 Find the error E in the answer E = xa^{2}
 If E is a positive number, a is too small.
If E is a negative number, a is too big.
 If the magnitude of E is sufficiently small, a = √x
 If the error E is too big modify a, return to step 2 and repeat.
The magnitude of E will progressively decrease until the desired accuracy is reached.
This method was designed for use with decimal numbers. It is not so convenient with binary numbers and can lead to an excessive number of loops to converge on an acceptable answer. Extra steps in this simple program can improve this.
 Direct Method
The square root can be calculated directly by using the properties of natural or base 10 logarithms but the logarithms themselves are transcendental functions and the values must first be extracted from a suitable approximation routine such as the one above.
The method depends on the following properties of the natural logarithm (ln):
 ln X^{n}=n ln X
and
 e ^{ln X } = X
The answer is given by: √x = e^{(ln X)/2} using the natural logarithm (ln)
Alternatively, using the base 10 logarithm (log), the answer is given by: √x = 10 e^{(log X)/2}
(Dividing the exponent by 2 generates a square root, the same as with logarithmic tables).
Pocket calculators typically use this method of calculating square roots.
CORDIC Approximation Algorithms
The CORDIC algorithm is an iterative method for calculating transcendental functions using only binary shift and add operations.
Using the example of calculating the value of a tangent, X/Y = tan Θ, expressed in the form tan(2^{n})
A vector from the origin (zero) is rotated in a series of "n" small steps δΘ so that the sum of all the steps is equal to Θ. By accumulating the corresponding changes in the values of its orthogonal coordinates δX and δY at each step, the values of X and Y and hence the tangent can be calculated.
Described by J.E. Volder in 1959, the CORDIC algorithm was used in the first scientific handheld calculator (the HP35) and variations of this simple basic concept have since been developed and refined for approximating a wide variety of transcendental functions.
