Files
The following zip file contains starter code, including the sort implementations provided by our textbook.
This zip file has the following contents:
- Sorting algorithms:
-
- BasicSorts.java - Includes selection sort and insertion sort. (UNFINISHED. See Part 2 below.)
-
- MergeSort.java - Includes merge sort.
-
- QuickSort.java - Includes quicksort.
-
- MergeSortImproved.java (UNFINISHED. See Parts 1 & 2 below.)
-
- IntrospectiveSort.java (UNFINISHED. See Part 3 below.)
- An application for timing sorting algorithms:
-
- SortProfiler.java
-
- Sorter.java - Functional Interface for sorting methods.
-
- Generator.java - Functional Interface for generator methods.
-
- Generators.java - Static methods for generating test data.
Note that SortProfiler relies on the JOpt Simple library for handling command-line arguments. You will need to have jopt-simple-5.0.4.jar in your classpath to use the application.
Part 1 - Improved Merges
The merge algorithm described in our textbook consists of the following two stages:
- Copy all values from the two pre-sorted halves into a temporary array.
- Merge the values from the temporary array back into the original array.
If \(n \) is the combined size of the two sub-arrays being merged, this algorithm requires a temporary array of size \(n \), and requires \(n \) assignment operations in stage 1. The following alternative approach cuts both of those values from \(n \) to around \(n/2 \)
- Copy the values from the first pre-sorted half to a temporary array.
- Merge the values from the temporary array and the second merge-sorted half into the original array.
Here are some ASCII graphics illustrating the process:
________________________________________
...| 1 | 3 | 5 | 7 | 9 | 2 | 4 | 6 | 8 | 10 |...
----------------------------------------
start-^ mid-^ end-^
___________________
| | | | | |...
-------------------
0-^
________________________________________
...| 1 | 3 | 5 | 7 | 9 | 2 | 4 | 6 | 8 | 10 |...
----------------------------------------
start-^ mid-^ end-^
___________________
| 1 | 3 | 5 | 7 | 9 |...
-------------------
________________________________________
...| | | | | | 2 | 4 | 6 | 8 | 10 |...
----------------------------------------
___________________
| 1 | 3 | 5 | 7 | 9 |...
-------------------
i1-^
________________________________________
...| | | | | | 2 | 4 | 6 | 8 | 10 |...
----------------------------------------
curr-^ i2-^
___________________
| 1 | 3 | 5 | 7 | 9 |...
-------------------
i1-^
________________________________________
...| 1 | | | | | 2 | 4 | 6 | 8 | 10 |...
----------------------------------------
curr-^ i2-^
___________________
| 1 | 3 | 5 | 7 | 9 |...
-------------------
i1-^
________________________________________
...| 1 | 2 | | | | 2 | 4 | 6 | 8 | 10 |...
----------------------------------------
curr-^ i2-^
___________________
| 1 | 3 | 5 | 7 | 9 |...
-------------------
________________________________________
...| 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 |...
----------------------------------------
Use the sort profiling tool to evaluate your finished implementation. You
should not expect your improved merge sort to be dramatically faster
than the provided implementation. Since array assignments are very
fast relative to calls to compareTo
, the speed impact will be
minimal. The main advantage here is the savings in space.
Part 2 - Switching Strategies
The next improvement is based on the observation that merge sort is actually slower than simple \(\Theta(n^2) \) sorts for small input sizes. This may seem surprising given that merge sort is an \(\Theta(nlogn) \) sorting algorithm. However, it is important to keep in mind that asymptotic analysis is only concerned with rates of growth. A \(\Theta(nlogn) \) algorithm will always be faster than a \(\Theta(n^2) \) algorithm eventually, but that doesn’t mean the \(\Theta(n^2) \) algorithm can’t be faster for small inputs. The following figure was created by timing merge sort and insertion sort on small randomly ordered arrays from size 2 to size 150:
As you can see, insertion sort is faster until around \(n=100 \). At that point, merge sort becomes faster and it remains faster for all larger inputs.
A a reminder, the following pseudocode describes the overall logic of the merge sort Algorithm:
merge_sort(sub-array)
If sub-array is has more than one entry:
Recursively merge_sort the left half
Recursively merge_sort the right half
Merge the two sorted halves.
This logic recursively splits the original array into smaller and smaller sub-arrays until the recursion bottoms out at sub-arrays of size one. This means that every time a large array is sorted, there are many recursive calls to merge sort that have small input sizes. In light of the figure above, that approach doesn’t make much sense: merge sort is not a competitive sorting algorithm on small inputs. It would make more sense to recursively break the input into smaller and smaller pieces until some threshold is reached, and then switch strategies to a sorting algorithm that is more efficient on those small inputs.
The following pseudocode describes this alternate approach:
merge_sort(sub-array)
If sub-array has fewer than MERGE_SORT_THRESHOLD entries:
Sort the sub-array with insertion sort.
Otherwise:
Recursively merge_sort the left half
Recursively merge_sort the right half
Merge the two sorted halves.
Choosing an appropriate value for MERGE_SORT_THRESHOLD
requires some
experimentation. One of the requirements of this assignment is that
you select an appropriate threshold value and provide data to justify
your choice. Simply selecting the crossover point based on the figure
above is not a reasonable strategy. The point where the two lines
cross represents the input size where the two sorts are equally
fast. Switching strategies at that input size will result in no
speedup at all.
The only way to determine an appropriate choice for
MERGE_SORT_THRESHOLD
is to systematically experiment with different
values until you find the one that leads to the best overall sorting
performance. You should create a driver method that iterates over
different values of MERGE_SORT_THRESHOLD
and uses SortProfiler
to
evaluate each. The following minimal driver class illustrates how
SortProfiler
objects can be used directly without invoking the
command line application:
import java.util.List;
public class ProfileDriver {
public static void main(String[] args) {
// Create a SortProfiler object.
// See the JavaDoc for an explanation of the parameters.
SortProfiler sp = new SortProfiler(List.of(MergeSort::mergeSort),
List.of("mergesort"),
1, 10, 100, 100,
Generators::generateRandom);
// Run the profiler and send the output to stdout.
sp.run(System.out);
}
}
Finally, note that we have provided an implementation of insertion sort that works on a full array, but you will need a version that works on a sub-array for this part of the assignment.
Part 3 - Introspective Sort
As mentioned above, quicksort is usually very fast. There are relatively few pathological inputs that result in \(\Theta(n^2) \) performance. The idea behind (introspective sort)[https://en.wikipedia.org/wiki/Introsort] is to perform a standard quick sort until there is evidence that the current input is pathological. If it looks like the current input will cause \(\Theta(n^2) \) performance, the algorithm switches strategies to a sorting algorithm with guaranteed \(\Theta(n logn) \) performance. Typically, heap sort is used as the alternative, since it is in-place and takes \(\Theta(nlogn) \) time in the worst case. Since we haven’t yet studied heap sort, your version of introspective sort should use the improved merge sort from Part 2 as the alternate sorting algorithm. The resulting algorithm is almost as fast as quick sort for most inputs, but has a worst case performance of \(\Theta(nlogn) \).
It is possible to detect pathological inputs by tracking the recursion depth of the quicksort algorithm. When quicksort is working well, the partition operation typically moves the pivot item somewhere near the center of the current sub-sequence. When this is the case, the maximum recursion depth will be \(logn \). Therefore, introspective sort switches strategies when the recursion depth exceeds \(2\lfloor log_2n \rfloor \).
There are a number of possible variations on the basic Introsort algorithm. Your implementation should conform to the following pseudocode:
introspective_sort(array)
introspective_sort(array, 2 ⌊log₂(array.length)⌋)
introspective_sort(sub-array, depth-limit)
If depth-limit is 0
merge_sort(sub-array)
Otherwise
If sub-array has 0 or 1 entries
return
Otherwise
pivot ← partition(sub-array)
introspective_sort(sub-array[0 to pivot - 1],
depth-limit - 1)
introspective_sort(sub-array[pivot + 1 to end],
depth-limit - 1)
Note that the provided version of merge sort sorts a full array. Introspective sort needs to be able to call a version of merge sort that can sort a portion of a provided array. The MergeSortImproved class provides a declaration for such a method.
Part 4 - Experimental Analysis
So far in this course we have focused on asymptotic analysis. This usually gives us the information we want. When deciding between two algorithms we generally prefer the one with better asymptotic performance.
Our goals for this project are somewhat different. The focus now is on fine-tuning existing algorithms to improve their performance. This is a situation where we really care about the constant factors. (We would be very happy if we could write a version of merge sort that is twice as fast, even if the asymptotic worst-case performance were still \(\Theta(nlogn) \).)
In order to help you evaluate your sorting improvements we are providing a command-line tool that can be used to systematically evaluate the run-time performance of different sorting algorithms.
Running the tool with no arguments should produce the following help message:
Option (* = required) Description
--------------------- -----------
* -s <Integer: NUMBER> Starting (smallest) input size
* -i <Integer: NUMBER> Input size increment
* -m <Integer: NUMBER> Maximum input size to test
* -t <Integer: NUMBER> Number of trials for each input size
-w [String: SORT1,SORT2,...] Comma separated list of sorts. Options include
insertion, selection, merge, merge_half,
merge_adaptive, quick, introspective and
timsort. Default is to execute all sorts.
-g [String: GENERATOR] Sequence generator. Options include random,
ordered or evil. The default is random
For example, we could compare the performance of quicksort and introspective sort on small inputs by executing the tool as follows:
$ java SortProfiler -s 1000 -i 1000 -m 10000 -t 10 -w quick,introspective
The resulting output would look something like the following:
N, quick, introspective
1000, 0.00009277, 0.00000085
2000, 0.00021152, 0.00000086
3000, 0.00030850, 0.00000088
4000, 0.00043710, 0.00000090
5000, 0.00055941, 0.00000088
6000, 0.00068144, 0.00000087
7000, 0.00081457, 0.00000091
8000, 0.00095075, 0.00000087
9000, 0.00105827, 0.00000089
10000, 0.00117701, 0.00000087
(Introspective sort is very fast here because it hasn’t been implemented yet.)
For this part of the project you will prepare a short document that quantifies the performance improvements you were able to achieve. This document should take the form of the following figures:
- Figure 1 - Improved Merges. This figure should compare the
performance of
mergeSortHalfSpace
to the original merge sort implementation as well as quicksort. Results should be shown for both random and ordered inputs. To avoid clutter, it may be helpful to plot the results for random inputs and ordered inputs separately. - Figure 2 - Threshold Selection. This figure should show the data
you used to select an appropriate threshold for switching
strategies in
mergeSortAdaptive
. - Figure 3 - Switching Strategies. This figure should compare your tuned
mergeSortAdaptive
to the previous two merge sort implementations as well as quicksort. Results should be shown for both random and ordered inputs. - Figure 4 - Introspective Sort (Random Inputs). This figure should compare introspective sort to quicksort for randomly generated inputs.
- Figure 5 - Introspective Sort (Pathological Inputs). This figure should compare introspective sort to quicksort for worst-case quicksort inputs.
Each figure must have clearly labeled axes, a caption, and appropriate legends. There must be a paragraph describing the results illustrated in each figure. The description must provide enough information for the reader to understand the point of the experiment and how the data supports the conclusion.
Submission
Submit the following five files through Gradescope:
BasicSorts.java
MergeSortImproved.java
IntrospectiveSort.java
SortAnalysis.pdf
- The document containing your analysisWhateverYouChooseToCallIt.java
- The driver class that you used to selectMERGE_SORT_THRESHOLD
.
Test your code locally at each stage of development, and reason about its correctness. When you are reasonably confident in your solution, submit your code to Gradescope to run the autograder’s unit tests. There will be a point deduction for excessive Gradescope submissions.
Grading
Gradescope Tests for Improved Merge | 20 points |
Gradescope Tests for Switching Strategies | 20 points |
Gradescope Tests for Introspective Sort | 20 points |
Experimental Results Document | 25 points |
Instructor Style Points | 10 points |
Gradescope Style Checks | 5 points |
Excessive Submission Deduction | -1 for each submission beyond 10. |
Any submission that doesn’t pass all of the instructor unit tests will receive at most 75% of the points for that component of the grade (45/60).
The grading of your evaluation document will be based on the quality of the figures as well as the clarity of your writing.
*The second approach has recently become popular. Both Python (starting in version 2.3) and Java (starting in version 7) use Timsort as their default sorting algorithm. Timsort is a modified version of merge sort that includes several enhancements: it uses a more space-efficient merge operation, it takes advantage of partially sorted arrays by finding and merging existing sorted regions, and it uses a non-recursive binary insertion sort to handle short unsorted regions.