The maximum element in an array can be looked up using a simple piece of code such as this piece of Javascript code. Given an input array A of size n:
var
M = A[ 0 ];
for
(
var
i = 0; i < n; ++i ) {
if
( A[ i ] >= M ) {
M = A[ i ];
}
}
- Assigning a value to a variable
- Looking up the value of a particular element in an array
- Comparing two values
- Incrementing a value
- Basic arithmetic operations such as addition and multiplication
if
and else
parts of code after the if
condition has been evaluated) occurs instantly and won't count these
instructions. In the above code, the first line of code is:var
M = A[ 0 ];
for
loop initialization code also has to always run. This gives us two more instructions; an assignment and a comparison:i = 0;
i < n;
for
loop iteration. After each for
loop iteration, we need two more instructions to run, an increment of i and a comparison to check if we'll stay in the loop:++i;
i < n;
for
loop and 2 instructions at the end of each iteration of which we have n. We can now define a mathematical function f( n ) that, given an n, gives us the number of instructions the algorithm needs. For an empty for
body, we have f( n ) = 4 + 2n.Worst-case analysis
Now, looking at thefor
body, we have an array lookup operation and a comparison that happen always:if
( A[ i ] >= M ) { ...
if
body may run or may not run, depending on what the array values actually are. If it happens to be so that A[ i ] >= M
, then we'll run these two additional instructions — an array lookup and an assignment:M = A[ i ]
A = [ 1, 2, 3, 4 ]
the algorithm will need more instructions than for A = [ 4, 3, 2, 1 ]
.
When analyzing algorithms, we often consider the worst-case scenario.
What's the worst that can happen for our algorithm? When does our
algorithm need the most instructions to complete? In this case, it is
when we have an array in increasing order such as A = [ 1, 2, 3, 4 ]
. In that case, M
needs to be replaced every single time and so that yields the most
instructions. Computer scientists have a fancy name for that and they
call it worst-case analysis; that's nothing more than just
considering the case when we're the most unlucky. So, in the worst case,
we have 4 instructions to run within the for
body, so we
have f( n ) = 4 + 2n + 4n = 6n + 4. This function f, given a problem
size n, gives us the number of instructions that would be needed in the
worst-case.Asymptotic behavior
Given such a function, we have a pretty good idea of how fast an algorithm is. However, as I promised, we won't be needing to go through the tedious task of counting instructions in our program. Besides, the number of actual CPU instructions needed for each programming language statement depends on the compiler of our programming language and on the available CPU instruction set (i.e. whether it's an AMD or an Intel Pentium on your PC, or a MIPS processor on your Playstation 2) and we said we'd be ignoring that. We'll now run our "f" function through a "filter" which will help us get rid of those minor details that computer scientists prefer to ignore.In our function, 6n + 4, we have two terms: 6n and 4. In complexity analysis we only care about what happens to the instruction-counting function as the program input (n) grows large. This really goes along with the previous ideas of "worst-case scenario" behavior: We're interested in how our algorithm behaves when treated badly; when it's challenged to do something hard. Notice that this is really useful when comparing algorithms. If an algorithm beats another algorithm for a large input, it's most probably true that the faster algorithm remains faster when given an easier, smaller input. From the terms that we are considering, we'll drop all the terms that grow slowly and only keep the ones that grow fast as n becomes larger. Clearly 4 remains a 4 as n grows larger, but 6n grows larger and larger, so it tends to matter more and more for larger problems. Therefore, the first thing we will do is drop the 4 and keep the function as f( n ) = 6n.
This makes sense if you think about it, as the 4 is simply an "initialization constant". Different programming languages may require a different time to set up. For example, Java needs some time to initialize its virtual machine. Since we're ignoring programming language differences, it only makes sense to ignore this value.
The second thing we'll ignore is the constant multiplier in front of n, and so our function will become f( n ) = n. As you can see this simplifies things quite a lot. Again, it makes some sense to drop this multiplicative constant if we think about how different programming languages compile. The "array lookup" statement in one language may compile to different instructions in different programming languages. For example, in C, doing
A[ i ]
does not include a check that i is within the declared array size, while in Pascal it does. So, the following Pascal code:M := A[ i ]
if
( i >= 0 && i < n ) {
M = A[ i ];
}
This filter of "dropping all factors" and of "keeping the largest growing term" as described above is what we call asymptotic behavior. So the asymptotic behavior of f( n ) = 2n + 8 is described by the function f( n ) = n. Mathematically speaking, what we're saying here is that we're interested in the limit of function f as n tends to infinity; but if you don't understand what that phrase formally means, don't worry, because this is all you need to know. (On a side note, in a strict mathematical setting, we would not be able to drop the constants in the limit; but for computer science purposes, we want to do that for the reasons described above.) Let's work a couple of examples to familiarize ourselves with the concept.
Let us find the asymptotic behavior of the following example functions by dropping the constant factors and by keeping the terms that grow the fastest.
- f( n ) = 5n + 12 gives f( n ) = n.
By using the exact same reasoning as above. - f( n ) = 109 gives f( n ) = 1.
We're dropping the multiplier 109 * 1, but we still have to put a 1 here to indicate that this function has a non-zero value. - f( n ) = n2 + 3n + 112 gives f( n ) = n2
Here, n2 grows larger than 3n for sufficiently large n, so we're keeping that. - f( n ) = n3 + 1999n + 1337 gives f( n ) = n3
Even though the factor in front of n is quite large, we can still find a large enough n so that n3 is bigger than 1999n. As we're interested in the behavior for very large values of n, we only keep n3 (See Figure 2). - f( n ) = n + √n gives f( n ) = n
This is so because n grows faster than √n as we increase n.
Exercise 1
- f( n ) = n6 + 3n
- f( n ) = 2n + 12
- f( n ) = 3n + 2n
- f( n ) = nn + n
If you're having trouble with one of the above, plug in some large n and see which term is bigger. Pretty straightforward, huh?
Complexity
So what this is telling us is that since we can drop all these decorative constants, it's pretty easy to tell the asymptotic behavior of the instruction-counting function of a program. In fact, any program that doesn't have any loops will have f( n ) = 1, since the number of instructions it needs is just a constant (unless it uses recursion; see below). Any program with a single loop which goes from 1 to n will have f( n ) = n, since it will do a constant number of instructions before the loop, a constant number of instructions after the loop, and a constant number of instructions within the loop which all run n times.This should now be much easier and less tedious than counting individual instructions, so let's take a look at a couple of examples to get familiar with this. The following PHP program checks to see if a particular value exists within an array A of size n:
$exists
= false;
for
(
$i
= 0;
$i
< n; ++
$i
) {
if
(
$A
[
$i
] ==
$value
) {
$exists
= true;
break
;
}
}
?>
Exercise 2
Systematically analyze the number of instructions the above PHP program needs with respect to n in the worst-case to find f( n ), similarly to how we analyzed our first Javascript program. Then verify that, asymptotically, we have f( n ) = n.v
=
a[
0
]
+
a[
1
]
The following program in C++ checks to see if a vector (a fancy array) named A of size n contains the same two values anywhere within it:
bool
duplicate =
false
;
for
(
int
i = 0; i < n; ++i ) {
for
(
int
j = 0; j < n; ++j ) {
if
( i != j && A[ i ] == A[ j ] ) {
duplicate =
true
;
break
;
}
}
if
( duplicate ) {
break
;
}
}
Rule of thumb: Simple programs can be analyzed by counting the nested loops of the program. A single loop over n items yields f( n ) = n. A loop within a loop yields f( n ) = n2. A loop within a loop within a loop yields f( n ) = n3.
If we have a program that calls a function within a loop and we know the number of instructions the called function performs, it's easy to determine the number of instructions of the whole program. Indeed, let's take a look at this C example:
int
i;
for
( i = 0; i < n; ++i ) {
f( n );
}
f( n )
is a function that performs exactly n instructions, we can then know that the number of instructions of the whole program is asymptotically n2, as the function is called exactly n times.
Rule of thumb: Given a
series of for loops that are sequential, the slowest of them determines
the asymptotic behavior of the program. Two nested loops followed by a
single loop is asymptotically the same as the nested loops alone,
because the nested loops dominate the simple loop.
- n6 + 3n ∈ Θ( n6 )
- 2n + 12 ∈ Θ( 2n )
- 3n + 2n ∈ Θ( 3n )
- nn + n ∈ Θ( nn )
We call this function, i.e. what we put within Θ( here ), the time complexity or just complexity of our algorithm. So an algorithm with Θ( n ) is of complexity n. We also have special names for Θ( 1 ), Θ( n ), Θ( n2 ) and Θ( log( n ) ) because they occur very often. We say that a Θ( 1 ) algorithm is a constant-time algorithm, Θ( n ) is linear, Θ( n2 ) is quadratic and Θ( log( n ) ) is logarithmic (don't worry if you don't know what logarithms are yet – we'll get to that in a minute).
Rule of thumb: Programs with a bigger Θ run slower than programs with a smaller Θ.
Big-O notation
Now, it's sometimes true that it will be hard to figure out exactly the behavior of an algorithm in this fashion as we did above, especially for more complex examples. However, we will be able to say that the behavior of our algorithm will never exceed a certain bound. This will make life easier for us, as we won't have to specify exactly how fast our algorithm runs, even when ignoring constants the way we did before. All we'll have to do is find a certain bound. This is explained easily with an example.A famous problem computer scientists use for teaching algorithms is the sorting problem. In the sorting problem, an array A of size n is given (sounds familiar?) and we are asked to write a program that sorts this array. This problem is interesting because it is a pragmatic problem in real systems. For example, a file explorer needs to sort the files it displays by name so that the user can navigate them with ease. Or, as another example, a video game may need to sort the 3D objects displayed in the world based on their distance from the player's eye inside the virtual world in order to determine what is visible and what isn't, something called the Visibility Problem (see Figure 3). The objects that turn out to be closest to the player are those visible, while those that are further may get hidden by the objects in front of them. Sorting is also interesting because there are many algorithms to solve it, some of which are worse than others. It's also an easy problem to define and to explain. So let's write a piece of code that sorts an array.
Here is an inefficient way to implement sorting an array in Ruby. (Of course, Ruby supports sorting arrays using build-in functions which you should use instead, and which are certainly faster than what we'll see here. But this is here for illustration purposes.)
b = []
n.times
do
m = a[
0
]
mi =
0
a.each_with_index
do
|element, i|
if
element < m
m = element
mi = i
end
end
a.delete_at( mi )
b << m
end
n - 1
times, then n - 2
times and so forth, until the last iteration of the outer loop during which it only runs once.It's a little harder to evaluate the complexity of this program, as we'd have to figure out the sum 1 + 2 + ... + (n - 1) + n. But we can for sure find an "upper bound" for it. That is, we can alter our program (you can do that in your mind, not in the actual code) to make it worse than it is and then find the complexity of that new program that we derived. If we can find the complexity of the worse program that we've constructed, then we know that our original program is at most that bad, or maybe better. That way, if we find out a pretty good complexity for our altered program, which is worse than our original, we can know that our original program will have a pretty good complexity too – either as good as our altered program or even better.
Let's now think of the way to edit this example program to make it easier to figure out its complexity. But let's keep in mind that we can only make it worse, i.e. make it take up more instructions, so that our estimate is meaningful for our original program. Clearly we can alter the inner loop of the program to always repeat exactly n times instead of a varying number of times. Some of these repetitions will be useless, but it will help us analyze the complexity of the resulting algorithm. If we make this simple change, then the new algorithm that we've constructed is clearly Θ( n2 ), because we have two nested loops where each repeats exactly n times. If that is so, we say that the original algorithm is O( n2 ). O( n2 ) is pronounced "big oh of n squared". What this says is that our program is asymptotically no worse than n2. It may even be better than that, or it may be the same as that. By the way, if our program is indeed Θ( n2 ), we can still say that it's O( n2 ). To help you realize that, imagine altering the original program in a way that doesn't change it much, but still makes it a little worse, such as adding a meaningless instruction at the beginning of the program. Doing this will alter the instruction-counting function by a simple constant, which is ignored when it comes to asymptotic behavior. So a program that is Θ( n2 ) is also O( n2 ).
But a program that is O( n2 ) may not be Θ( n2 ). For example, any program that is Θ( n ) is also O( n2 ) in addition to being O( n ). If we imagine the that a Θ( n ) program is a simple
for
loop that repeats n times, we can make it worse by wrapping it in another for
loop which repeats n times as well, thus producing a program with f( n ) = n2. To generalize this, any program that is Θ( a ) is O( b ) when b is worse than a.
Notice that our alteration to the program doesn't need to give us a
program that is actually meaningful or equivalent to our original
program. It only needs to perform more instructions than the original
for a given n. All we're using it for is counting instructions, not actually solving our problem.So, saying that our program is O( n2 ) is being on the safe side: We've analyzed our algorithm, and we've found that it's never worse than n2. But it could be that it's in fact n2. This gives us a good estimate of how fast our program runs.
Rule of thumb: It's easier to figure out the O-complexity of an algorithm than its Θ-complexity.
The reason we use O and Ω instead of Θ even though O and Ω can also give tight bounds is that we may not be able to tell if a bound we've found is tight, or we may just not want to go through the process of scrutinizing it so much.
If you don't fully remember all the different symbols and their uses, don't worry about it too much right now. You can always come back and look them up. The most important symbols are O and Θ.
Also note that although Ω gives us a lower-bound behavior for our function (i.e. we've improved our program and made it perform less instructions) we're still referring to a "worst-case" analysis. This is because we're feeding our program the worst possible input for a given n and analyzing its behavior under this assumption.
The following table indicates the symbols we just introduced and their correspondence with the usual mathematical symbols of comparisons that we use for numbers. The reason we don't use the usual symbols here and use Greek letters instead is to point out that we're doing an asymptotic behavior comparison, not just a simple comparison.
Asymptotic comparison operator | Numeric comparison operator |
---|---|
Our algorithm is o( something ) | A number is < something |
Our algorithm is O( something ) | A number is ≤ something |
Our algorithm is Θ( something ) | A number is = something |
Our algorithm is Ω( something ) | A number is ≥ something |
Our algorithm is ω( something ) | A number is > something |
Rule of thumb: While all
the symbols O, o, Ω, ω and Θ are useful at times, O is the one used more
commonly, as it's easier to determine than Θ and more practically
useful than Ω.
Logarithms
Logarithms are important because they occur a lot when analyzing complexity. A logarithm is an operation applied to a number that makes it quite smaller – much like a square root of a number. So if there's one thing you want to remember about logarithms is that they take a number and make it much smaller than the original. Now, in the same way that square roots are the inverse operation of squaring something, logarithms are the inverse operation of exponentiating something. This isn't as hard as it sounds. It's better explained with an example. Consider the equation:2x = 1024
We now wish to solve this equation for x. So we ask ourselves: What is the number to which we must raise the base 2 so that we get 1024? That number is 10. Indeed, we have 210 = 1024, which is easy to verify. Logarithms help us denote this problem using new notation. In this case, 10 is the logarithm of 1024 and we write this as log( 1024 ) and we read it as "the logarithm of 1024". Because we're using 2 as a base, these logarithms are called base 2 logarithms. There are logarithms in other bases, but we'll only use base 2 logarithms in this article. If you're a student competing in international competitions and you don't know about logarithms, I highly recommend that you practice your logarithms after completing this article. In computer science, base 2 logarithms are much more common than any other types of logarithms. This is because we often only have two different entities: 0 and 1. We also tend to cut down one big problem into halves, of which there are always two. So you only need to know about base-2 logarithms to continue with this article.
Rule of thumb: For computing algorithms implemented in C++, once you've analyzed your complexity, you can get a rough estimate of how fast your program will run by expecting it to perform about 1,000,000 operations per second, where the operations you count are given by the asymptotic behavior function describing your algorithm. For example, a Θ( n ) algorithm takes about a second to process the input for n = 1,000,000.
Recursive complexity
Let's now take a look at a recursive function. A recursive function is a function that calls itself. Can we analyze its complexity? The following function, written in Python, evaluates the factorial of a given number. The factorial of a positive integer number is found by multiplying it with all the previous positive integers together. For example, the factorial of 5 is 5 * 4 * 3 * 2 * 1. We denote that "5!" and pronounce it "five factorial" (some people prefer to pronounce it by screaming it out aloud like "FIVE!!!")def
factorial( n ):
if
n
=
=
1
:
return
1
return
n
*
factorial( n
-
1
)
If you're unsure about this fact, remember that you can always find the exact complexity by counting instructions. If you wish, you can now try to count the actual instructions performed by this function to find a function f( n ) and see that it's indeed linear (recall that linear means Θ( n )).
Logarithmic complexity
One famous problem in computer science is that of searching for a value within an array. We solved this problem earlier for the general case. This problem becomes interesting if we have an array which is sorted and we want to find a given value within it. One method to do that is called binary search. We look at the middle element of our array: If we find it there, we're done. Otherwise, if the value we find there is bigger than the value we're looking for, we know that our element will be on the left part of the array. Otherwise, we know it'll be on the right part of the array. We can keep cutting these smaller arrays in halves until we have a single element to look at. Here's the method using pseudocode:def
binarySearch( A, n, value ):
if
n
=
1
:
if
A[
0
]
=
value:
return
true
else
:
return
false
if
value < A[ n
/
2
]:
return
binarySearch( A[
0.
..( n
/
2
-
1
) ], n
/
2
-
1
, value )
else
if
value > A[ n
/
2
]:
return
binarySearch( A[ ( n
/
2
+
1
)...n ], n
/
2
-
1
, value )
else
:
return
true
If you're unsure that this method actually works, take a moment now to run it by hand in a simple example and convince yourself that it actually works.
Let us now attempt to analyze this algorithm. Again, we have a recursive algorithm in this case. Let's assume, for simplicity, that the array is always cut in exactly a half, ignoring just now the + 1 and - 1 part in the recursive call. By now you should be convinced that a little change such as ignoring + 1 and - 1 won't affect our complexity results. This is a fact that we would normally have to prove if we wanted to be prudent from a mathematical point of view, but practically it is intuitively obvious. Let's assume that our array has a size that is an exact power of 2, for simplicity. Again this assumption doesn't change the final results of our complexity that we will arrive at. The worst-case scenario for this problem would happen when the value we're looking for does not occur in our array at all. In that case, we'd start with an array of size n in the first call of the recursion, then get an array of size n / 2 in the next call. Then we'll get an array of size n / 4 in the next recursive call, followed by an array of size n / 8 and so forth. In general, our array is split in half in every call, until we reach 1. So, let's write the number of elements in our array for every call:
Notice that in the i-th iteration, our array has n / 2i elements. This is because in every iteration we're cutting our array into half, meaning we're dividing its number of elements by two. This translates to multiplying the denominator with a 2. If we do that i times, we get n / 2i. Now, this procedure continues and with every larger i we get a smaller number of elements until we reach the last iteration in which we have only 1 element left. If we wish to find i to see in what iteration this will take place, we have to solve the following equation:
1 = n / 2i
This will only be true when we have reached the final call to the binarySearch() function, not in the general case. So solving for i here will help us find in which iteration the recursion will finish. Multiplying both sides by 2i we get:
2i = n
Now, this equation should look familiar if you read the logarithms section above. Solving for i we have:
i = log( n )
This tells us that the number of iterations required to perform a binary search is log( n ) where n is the number of elements in the original array.
If you think about it, this makes some sense. For example, take n = 32, an array of 32 elements. How many times do we have to cut this in half to get only 1 element? We get: 32 → 16 → 8 → 4 → 2 → 1. We did this 5 times, which is the logarithm of 32. Therefore, the complexity of binary search is Θ( log( n ) ).
This last result allows us to compare binary search with linear search, our previous method. Clearly, as log( n ) is much smaller than n, it is reasonable to conclude that binary search is a much faster method to search within an array then linear search, so it may be advisable to keep our arrays sorted if we want to do many searches within them.
Rule of thumb: Improving
the asymptotic running time of a program often tremendously increases
its performance, much more than any smaller "technical" optimizations
such as using a faster programming language.
Optimal sorting
You now know about analyzing the complexity of algorithms, asymptotic behavior of functions and big-O notation. You also know how to intuitively figure out that the complexity of an algorithm is O( 1 ), O( log( n ) ), O( n ), O( n2 ) and so forth. You know the symbols o, O, ω, Ω and Θ and what worst-case analysis means. If you've come this far, this tutorial has already served its purpose.This final section is optional. It is a little more involved, so feel free to skip it if you feel overwhelmed by it. It will require you to focus and spend some moments working through the exercises. However, it will provide you with a very useful method in algorithm complexity analysis which can be very powerful, so it's certainly worth understanding.
We looked at a sorting implementation above called a selection sort. We mentioned that selection sort is not optimal. An optimal algorithm is an algorithm that solves a problem in the best possible way, meaning there are no better algorithms for this. This means that all other algorithms for solving the problem have a worse or equal complexity to that optimal algorithm. There may be many optimal algorithms for a problem that all share the same complexity. The sorting problem can be solved optimally in various ways. We can use the same idea as with binary search to sort quickly. This sorting method is called mergesort.
To perform a mergesort, we will first need to build a helper function that we will then use to do the actual sorting. We will make a
merge
function which takes two arrays that are both already sorted and merges
them together into a big sorted array. This is easily done:def
merge( A, B ):
if
empty( A ):
return
B
if
empty( B ):
return
A
if
A[
0
] < B[
0
]:
return
concat( A[
0
], merge( A[
1.
..A_n ], B ) )
else
:
return
concat( B[
0
], merge( A, B[
1.
..B_n ] ) )
def
mergeSort( A, n ):
if
n
=
1
:
return
A
# it is already sorted
middle
=
floor( n
/
2
)
leftHalf
=
A[
1.
..middle ]
rightHalf
=
A[ ( middle
+
1
)...n ]
return
merge( mergeSort( leftHalf, middle ), mergeSort( rightHalf, n
-
middle ) )
As a final example, let us analyze the complexity of
mergeSort
. In every step of mergeSort
, we're splitting the array into two halves of equal size, similarly to binarySearch
.
However, in this case, we maintain both halves throughout execution. We
then apply the algorithm recursively in each half. After the recursion
returns, we apply the merge
operation on the result which takes Θ( n ) time.So, we split the original array into two arrays of size n / 2 each. Then we merge those arrays, an operation that merges n elements and thus takes Θ( n ) time.
Take a look at the Figure to understand this recursion.
Let's see what's going on here. Each circle represents a call to the
mergeSort
function. The number written in the circle indicates the size of the
array that is being sorted. The top blue circle is the original call to mergeSort
, where we get to sort an array of size n. The arrows indicate recursive calls made between functions. The original call to mergeSort
makes two calls to mergeSort
on two arrays, each of size n / 2. This is indicated by the two arrows
at the top. In turn, each of these calls makes two calls of its own to mergeSort
two arrays of size n / 4 each, and so forth until we arrive at arrays of size 1. This diagram is called a recursion tree, because it illustrates how the recursion behaves and looks like a tree (the root is at the top and the leaves are at the bottom, so in reality it looks like an inversed tree).Notice that at each row in the above diagram, the total number of elements is n. To see this, take a look at each row individually. The first row contains only one call to
mergeSort
with an array of size n, so the total number of elements is n. The second row has two calls to mergeSort
each of size n / 2. But n / 2 + n / 2 = n and so again in this row the total number of elements is n.
In the third row, we have 4 calls each of which is applied on an n /
4-sized array, yielding a total number of elements equal to n / 4 + n / 4
+ n / 4 + n / 4 = 4n / 4 = n. So again we get n elements. Now notice that at each row in this diagram the caller will have to perform a merge
operation on the elements returned by the callees. For example, the
circle indicated with red color has to sort n / 2 elements. To do this,
it splits the n / 2-sized array into two n / 4-sized arrays, calls mergeSort
recursively to sort those (these calls are the circles indicated with
green color), then merges them together. This merge operation requires
to merge n / 2 elements. At each row in our tree, the total number of
elements merged is n. In the row that we just explored, our function
merges n / 2 elements and the function on its right (which is in blue
color) also has to merge n / 2 elements of its own. That yields n
elements in total that need to be merged for the row we're looking at.By this argument, the complexity for each row is Θ( n ). We know that the number of rows in this diagram, also called the depth of the recursion tree, will be log( n ). The reasoning for this is exactly the same as the one we used when analyzing the complexity of binary search. We have log( n ) rows and each of them is Θ( n ), therefore the complexity of
mergeSort
is Θ( n * log( n ) ). This is much better than Θ( n2
) which is what selection sort gave us (remember that log( n ) is much
smaller than n, and so n * log( n ) is much smaller than n * n = n2).
If this sounds complicated to you, don't worry: It's not easy the first
time you see it. Revisit this section and reread about the arguments
here after you implement mergesort in your favourite programming
language and validate that it works.As you saw in this last example, complexity analysis allows us to compare algorithms to see which one is better. Under these circumstances, we can now be pretty certain that merge sort will outperform selection sort for large arrays. This conclusion would be hard to draw if we didn't have the theoretical background of algorithm analysis that we developed. In practice, indeed sorting algorithms of running time Θ( n * log( n ) ) are used. For example, the Linux kernel uses a sorting algorithm called heapsort, which has the same running time as mergesort which we explored here, namely Θ( n log( n ) ) and so is optimal. Notice that we have not proven that these sorting algorithms are optimal. Doing this requires a slightly more involved mathematical argument, but rest assured that they can't get any better from a complexity point of view.
No comments:
Post a Comment