Simply put, recursion is when a function calls itself. That is, in the course of the function definition there is a call to that very same function. At first this may seem like a never ending loop, or like a dog chasing its tail. It can never catch it. So too it seems our function will never finish. This might be true is some cases, but in practice we can check to see if a certain condition is true and in that case exit (return from) our function. The case in which we end our recursion is called a base case . Additionally, just as in a loop, we must change some value and incremently advance closer to our base case.
Consider this function.
void myFunction( int counter)
{
if(counter == 0)
return;
else
{
cout <<counter<<endl;
myFunction(--counter);
return;
}
}
This recursion is not infinite, assuming the function is passed a
positive integer value. What will the output be?
Consider this function:
void myFunction( int counter)
{
if(counter == 0)
return;
else
{
cout<<"hello"<<counter<<endl;
myFunction(--counter);
cout<<counter<<endl;
return;
}
}
If the function is called with the value 4, what will the output be? Explain.
The above recursion is essentially a loop like a for loop or a while loop.
When do we prefer recursion to an iterative loop? We use recursion
when we can see that our problem can be reduced to
a simpler problem that can be solved after further reduction.
Every recursion should have the following characteristics.
The key to thinking recursively is to see the solution to the problem as a smaller version of the same problem. The key to solving recursive programming requirements is to imagine that your function does what its name says it does even before you have actually finish writing it. You must pretend the function does its job and then use it to solve the more complex cases. Here is how.
Identify the base case(s) and what the base case(s) do. A base case is the simplest possible problem (or case) your function could be passed. Return the correct value for the base case. Your recursive function will then be comprised of an if-else statement where the base case returns one value and the non-base case(s) recursively call(s) the same function with a smaller parameter or set of data. Thus you decompose your problem into two parts: (1) The simplest possible case which you can answer (and return for), and (2) all other more complex cases which you will solve by returning the result of a second calling of your function. This second calling of your function ( recursion ) will pass on the complex problem but reduced by one increment. This decomposition of the problem will actually be a complete, accurate solution for the problem for all cases other than the base case. Thus, the code of the function actually has the solution on the first recursion.
Recursion is a lot like proof by mathematical induction. That is, you show that if a statement is true for any
number, n, then it must also be true for n+1. Or we could show that if the statement is true
for n-1 then it must be true for n. Then you show that it is true for n=0 (or n=1). So too, in solving a
a problem recursively, we write our function to work for n while we assume the function works for n-1.
we then make the function work in the base case, i.e. where n=1. The key is to assume our function works
when writing it. That is why we can call it from within itself. The logic is that if it works for n=1 (the base case
with an explicit return), then when solving the problem for n=2 we can assume it works for n=1 because
we have explicitly solved that problem. Now the solution for n=3 must also work, since
we have a solution for n=2 (i.e. n = 3-1). This thinking can go on ad infinitum and therefore
we have solved the problem for any n.
Let's consider writing a function to find the factorial of an integer.
For example 7! equals 7*6*5*4*3*2*1 .
But we are also correct if we say 7! equals 7*6!.
In seeing the factorial of 7 in this second way
we have gained a valuable insight. We now can see our problem in terms
of a simpler version of our problem and we even know
how to make our problem progressively more simple. We have also defined
our problem in terms of itself. I.e. we defined 7! in terms of 6!.
This is the essence of recursive problem solving.
Now all we have left to do is decide what the base case is.
What is the simplest factorial? 1!. 1! equals 1.
Let's write the factorial function recursively.
int myFactorial( int integer)
{
if( integer == 1)
return 1;
else
{
return (integer * (myFactorial(integer-1)));
}
}
Note that the base case ( the factorial of 1 ) is solved and the return value
is given. Now let us imagine that our function actually works. If it works
we can use it to give the result of more complex cases.
If our number is 7 we will simply return 7 * the result of factorial of 6.
So we actaully have the exact answer for all cases in the top level recursion.
Our problem is getting smaller on each recursive call because each time we
call the function we give it a smaller number.
Try running this program in your head with the number 2. Does it give the right
value? If it works for 1 then it must work for two since 2 merely
returns 2 * factorial of 1. Now will it work for 3? Well, 3 must return
3 * factorial of 2. Now since we know that factorial of 2 works, factorial
of 3 also works. We can prove that 4 works in the same way, and so on and so on.
Food for thought: ask yourself, could this be written iteratively?
Note: make it your habit of writing the base case in
the function as the first statement.
Note: Forgetting the base case leads to infinite recursion.
However, in fact, your code won't run forever like an infinite loop, instead,
you will eventually run out of stack space (memory) and get a run-time
error or exception called a stack overflow.
There are several significant problems with recursion.
Mostly it is hard (especially for inexperienced programmers) to think
recursively, though many AI specialists claim that in reality
recursion is closer to basic human thought processes
than other programming functions (such as iteration).
There also exists the problem of stack overflow when using some
forms of recursion (head recursion.)
The other main problem with recursion is that it can be slower to run
than simple iteration.
Then why use it? It seems that there is always an iterative solution to any
problem that can be solved recursively.
Is there a difference in computational complexity? No.
Is there a difference in the efficiency of execution? Yes, in fact, the
recursive version is usually less efficient because of having to push and
and pop recursions on and off the run-time stack, so iteration is quicker.
On the other hand, you might notice that the recursive versions use fewer or no
local variables.
So why use recursion? The answer to our question is predominantly because it is easier to code a recursive solution once one is able to identify that solution. The recursive code is usually smaller, more concise, more elegant, possibly even easier to understand, though that depends on ones thinking style. But also, there are some problems that are very difficult to solve without recursion. Those problems that require backtracking such as searching a maze for a path to an exit or tree based operations (which we will see in semester 2) are best solved recursively. There are also some interesting sorting algorithms that use recursion.
This problem comes from history, monks in Vietnam were asked to carry 64 gold disks from one tower (stack) to another. Each disk is of a different size. There are 3 stacks, a source stack, a destination stack and an intermediate stack. A disk is placed on one of three stacks but no disk can be placed on top of a smaller disk. The source tower holds 64 disks. How will the monks solve this problem? How long will it take them?
The easiest solution is a recursive one. The key to the solution is to notice that to move any disk, we must first move the smaller disks off of it, thus a recursive definition. Another way to look at it is this, if we had a function to move the top three disks to the middle position, we could put the biggest disk in its place. All we need to do is assume we have this function and then call it.
Lets start with 1 disk (our base case):
Move 1 disk from start tower to destination tower
and we are done.
To move 2 disks:
Move smaller disk from start tower to intermediate tower, move larger
disk from start tower to final tower, move smaller disk from
intermediate tower to final tower and we are done.
To move n disks (or think of, say, 3 disks):
Solve the problem for n - 1 disks (i.e. 2 disks) using the intermediate
tower instead of the final tower (i.e. get 2 disks onto the intermediate tower).
Then, move the biggest disk from start tower to final tower.
Then again solve the problem for n - 1 disks but use the intermediate
tower instead of the start tower (i.e. get the 2 disks onto the final tower using
the start tower as the intermediate tower).
Tail recursion is defined as occuring when the recursive call is at the end of the recursive instruction. This is not the case with my factorial solution above. It is useful to notice when ones algorithm uses tail recursion because in such a case, the algorithm can usually be rewritten to use iteration instead. In fact, the compiler will (or at least should) convert the recursive program into an iterative one. This eliminates the potential problem of stack overflow.
This is not the case with head recursion, or when the function calls itself recursively in different places like in the Towers of Hanoi solution. Of course, even in these cases we could also remove recursion by using our own stack and essentially simulating how recursion would work.
In my example of factorial above the compiler will have to call the recursive function before doing the multiplication because it has to resolve the (return) value of the function before it can complete the multiplication. So the order of execution will be "head" recursion, i.e. recursion occurs before other operations.
To convert this to tail recursion we need to get all the multiplication finished and resolved before recursively calling the function. We need to force the order of operation so that we are not waiting on multiplication before returning. If we do this the stack frame can be freed up.
The proper way to do a tail-recursive factorial is this:
int factorial(int number) { if(number == 0) { return 1; } return factorial_i(number, 1); } int factorial_i(int currentNumber, int sum) { if(currentNumber == 1) { return sum; } else { return factorial_i(currentNumber - 1, sum*currentNumber); } }Notice that in the call return factorial_i(currentNumber - 1, sum*currentNumber); both parameters are immediately resolvable. We can compute what each parameter is without waiting for a recursive function call to return. This is not the case with the previous version of factorial. This streamlining enables the compiler to minimize stack use as explained above. Thanks to Jon Bartlett for the example.
Some definitions ( types of recursion ):
p) return 1; if (p%i == 0) return 0; return isPrime (p, i+1); } // two versions of recursive solution to adding up numbers from 1 to any given number. // the second example is tail recusion because once the total is found, the function returns and // does not need to unravel previous recursive steps int sum (int num) { if (num==0) return 0; return (sum(num-1)+(num)); } int sum (int num, int total=0) { if (num<=0) return total; sum( num-1, sum ); }
© Nachum Danzig