Unsolved problem in computer science:
Is there an sorting algorithm faster than ?

In computer science, sorting is the problem of sorting pairs of numbers by their sums. Applications of the problem include transit fare minimisation, VLSI design, and sparse polynomial multiplication. As with comparison sorting and integer sorting more generally, algorithms for this problem can be based only on comparisons of these sums, or on other operations that work only when the inputs are small integers.

Geometric visualization of the sorting problem. The input sets and are represented by the sets of vertical and horizontal black lines (respectively), and the goal of the problem is to sort the crossing points by the positions of the red diagonal lines through them.

It is unknown whether this problem has a comparison-based solution whose running time is asymptotically faster than sorting an unstructured list of equally many items. Therefore, research on the problem has focused on two approaches to settle the question of whether such an improvement is possible: the development of algorithms that improve on unstructured sorting in their number of comparisons rather than in their total running time, and lower bounds for the number of comparisons based on counting cells in subdivisions of high-dimensional spaces. Both approaches are historically tied together, in that the first algorithms that used few comparisons were based on the weakness of the cell-counting lower bounds.

Problem statement and history

edit

The input to the   sorting problem consists of two finite collections of numbers   and  , of the same length. The problem's output is the collection of all pairs of a number from   and a number from  , arranged into sorted order by the sum of each pair.[1] As a small example, for the inputs   and  , the output should be the list of pairs of one element from   and one element from  , listed in sorted order by their sums of pairs One way to solve the problem would be to construct the pairs to be sorted (the Cartesian product of the two collections) and use these pairs as input to a standard comparison sorting algorithm such as merge sort or heapsort. When the inputs have length  , they form   pairs, and the time to sort the pairs in this way is  . In terms of its big O notation, this method is the fastest known algorithm for   sorting. Whether a faster algorithm exists is an open problem,[1][2] posed by Elwyn Berlekamp prior to 1975.[1][3]

A variant of the problem sorts the sumset, the set of sums of pairs, with duplicate sums condensed to a single value. For this variant, the size of the sumset may be significantly smaller than  , and output-sensitive algorithms for constructing it have been investigated.[4]

Applications

edit

Steven Skiena recounts a practical application in transit fare minimisation, an instance of the shortest path problem: find the cheapest two-hop airplane ticket between two given cities, from an input that describes both the cost of each hop and which pairs of hops may be combined into a single ticket. Skiena's solution consists of sorting pairs of hops by their total cost as an instance of the   sorting problem, and then testing the resulting pairs in this sorted order until finding one that is allowed. To generate the sorted pairs in this order, Skiena uses a priority queue of pairs, initially containing only a single pair, the one consisting of the two cheapest hops. Then, when a pair   is removed from the queue and found to be disallowed, two more pairs are added, with one of these two pairs combining   with the next hop after   in a sorted list of the hops to the destination, and the other pair combining   with the next hop after   in a sorted list of hops from the start. In this way, each successive pair can be found in logarithmic time, and only the pairs up to the first allowable one need to be sorted.[2]

  sorting is the most expensive subroutine in an algorithm for a problem in VLSI design, in which one must place two subunits of a VLSI circuit side-by-side along a communications channel in order to minimize the width of the channel needed to route pairs of wires from one subunit to the other. As one subunit is continuously shifted relative to the other, the channel width only changes at discrete positions where the ends of two wires line up with each other, and finding the sorted ordering of these positions in order to compute the sequence of changes to the width can be performed by   sorting. If this sorting problem could be sped up, it would also speed up this VLSI design task.[5]

Another application involves polynomial multiplication for polynomials of a single variable that may have many fewer terms than their degrees. The product of two polynomials can be expressed as a sum of products of pairs of terms, one from each polynomial, and placing these term-by-term products into degree order amounts to sorting them by the sum of degrees. For example, the instance of   sorting with   given as an example above corresponds to the multiplication of two three-term polynomials to produce a nine-term polynomial:   The degrees are always integers, so integer-based algorithms for   sorting may be applied.[6] However, for polynomials whose number of terms is comparable to their degree, FFT-based polynomial multiplication algorithms may be significantly more efficient than term-by-term multiplication.[7]

Number of orderings

edit

A well-known lower bound for unstructured sorting, in the decision tree model, is based on the factorial number of sorted orders that an unstructured list may have. Because each comparison can at best reduce the number of possible orderings by a factor of two, sorting requires a number of comparisons at least equal to the binary logarithm of the factorial, which is  .[8] Early work on   sorting followed a similar approach by asking how many different sorted orderings are possible for this problem, and proving that this number is at most  . However, because its binary logarithm is at most  , much smaller than the known time bounds for   sorting, this method can only lead to weak lower bounds on the number of comparisons.[3][9]

The proof of this bound relates   sorting to the complexity of an arrangement of hyperplanes in high-dimensional geometry. The two input collections for the   sorting problem comprise   numbers, which can alternatively be interpreted as the Cartesian coordinates of a point in the  -dimensional space  . This space can be subdivided into cells, so that within a single cell all points correspond to inputs that produce the same sorted order. For this subdivision, each boundary between two cells lies within a hyperplane defined by an equality of pairs  , where   and   are two pairs whose ordering changes from one adjacent cell to the other. These hyperplanes are either generated by two disjoint pairs, or they have the simplified forms   or  , so the number of distinct hyperplanes that can be determined in this way is   The number of cells that this number of hyperplanes can divide a space of dimension   into is   Therefore, the set   has   different possible sorted orderings.[3][9][10]

A similar style of analysis has been more successful in ruling out fast solutions to certain generalizations of   sorting, by showing that they have too many orderings to sort quickly. In particular, Harper et al. (1975) suggest separately sorting   and  , and then constructing a two-dimensional matrix of the values of   that is sorted both by rows and by columns before using this partially-sorted data to complete the sort of  .[3] This idea of using a row-sorted and column-sorted matrix forms the basis for the method used by Skiena in the transportation application,[2] and it can reduce the number of comparisons by a constant factor relative to naive comparison sorting. However, for matrices whose rows and columns are sorted in this way, the number of possible sorted orderings of the whole matrix is much larger than  , so large that any comparison sorting algorithm that can work for arbitrary   matrices that are sorted by rows and columns still requires   comparisons. Therefore, if the   sorting problem is to be solved quickly, the solution must use additional information about the set   beyond this matrix ordering.[3]

Number of comparisons

edit

For the classical comparison sorting problem, the time to sort and the number of comparisons needed to sort are within constant factors of each other. But for   sorting, the number of comparisons is smaller than the best time bound known: Michael Fredman showed in 1976 that   sorting can be done using only   comparisons. More generally, he showed that any set of   elements, whose sorted ordering has already been restricted to a family   of orderings, can be sorted using   comparisons, by a form of binary insertion sort. For the   sorting problem,  , and  , so   and Fredman's bound implies that only   comparisons are needed. However, in Fredman's method, the time needed to decide which comparisons to perform may be significantly higher than the bound on the number of comparisons.[9]

The first explicit algorithm that achieves both   comparisons and   total complexity was published sixteen years after Fredman by Lambert (1992). The algorithm performs the following steps:

  1. Recursively sort the two sets   and  .
  2. Use the equivalence   to infer the sorted orderings of   and   without additional comparisons.
  3. Merge the two sets   and   into a single sorted order, using a number of comparisons linear in their total size.
  4. Use the merged order and the equivalence   to infer the sorted order of   without additional comparisons.

The part of the algorithm that recursively sorts   (or equivalently  ) does so by the following steps:

  1. Split   into two equal sublists   and  .
  2. Recursively sort   and  
  3. Infer the ordering on   using only the comparisons from a single merge step as above.
  4. Merge the sorted results  ,  , and   together.

The number of comparisons   needed to perform this recursive algorithm on an input of   items can be analyzed using the recurrence relation   where the   term of the recurrence counts the number of comparisons in the recursive calls to the algorithm to sort   and  , and the   term counts the number of comparisons used to merge the results. The master theorem for recurrence relations of this form shows that   The total time complexity is slower,  , because of the steps of the algorithm that use already-made comparisons to infer orderings of other sets. These steps can be performed in time   by using a standard comparison-sorting algorithm with its comparison steps replaced by the stated inferences.[11]

If only comparisons between elements of   are allowed, then there is also a matching lower bound of   on the number of comparisons,[9][12] but with more general comparisons involving linear combinations of constant numbers of elements, only   comparisons are needed.[13]

Non-comparison-based algorithms

edit

Just as integer sorting can be faster than comparison sorting for small-enough integers, the same is true for   sorting. In particular, with integer inputs in the range from   to some upper limit  , the problem can be solved in   operations by means of the fast Fourier transform.[1][3]

edit

Several other problems in computational geometry have equivalent or harder complexity to   sorting, including constructing Minkowski sums of staircase polygons, finding the crossing points of an arrangement of lines in sorted order by their  -coordinates, listing pairs of points in sorted order by their distances, and testing whether one rectilinear polygon can be translated to fit within another.[14]

The problem of testing whether two of the pairs in the   sorting problem have equal sums can be solved by sorting the pairs and then testing consecutive pairs for equality. In turn, it could be used to solve the 3SUM problem, implying that it is unlikely to have a strongly subquadratic algorithm.[1]

References

edit
  1. ^ a b c d e Demaine, Erik; Erickson, Jeff; O'Rourke, Joseph (20 August 2006). "Problem 41: Sorting X + Y (Pairwise Sums)". The Open Problems Project. Retrieved 23 September 2014.
  2. ^ a b c Skiena, Steven (2008). "4.4 War Story: Give me a Ticket on an Airplane". The Algorithm Design Manual (2nd ed.). Springer. pp. 118–120. doi:10.1007/978-1-84800-070-4_4.
  3. ^ a b c d e f Harper, L. H.; Payne, T. H.; Savage, J. E.; Straus, E. (1975). "Sorting X + Y". Communications of the ACM. 18 (6): 347–349. doi:10.1145/360825.360869. MR 0378473. S2CID 26360885.
  4. ^ Arnold, Andrew; Roche, Daniel S. (2015). "Output-sensitive algorithms for sumset and sparse polynomial multiplication". Proceedings of the 2015 ACM International Symposium on Symbolic and Algebraic Computation (ISSAC'15). New York: ACM. pp. 29–36. arXiv:1501.05296. doi:10.1145/2755996.2756653. ISBN 978-1-4503-3435-8. MR 3388279.
  5. ^ LaPaugh, Andrea S.; Pinter, Ron Y. (1983). "On minimizing channel density by lateral shifting". IEEE International Conference on Computer-Aided Design. pp. 123–124. As cited by Johnson, David S.; LaPaugh, Andrea S.; Pinter, Ron Y. (1994). "Minimizing channel density by lateral shifting of components". In Sleator, Daniel Dominic (ed.). Proceedings of the Fifth Annual ACM-SIAM Symposium on Discrete Algorithms. 23-25 January 1994, Arlington, Virginia, USA. pp. 122–131.
  6. ^ Klip, Dorothea A. (1979). "New algorithms for polynomial multiplication". SIAM Journal on Computing. 8 (3): 326–343. doi:10.1137/0208025. MR 0539251.
  7. ^ Klarreich, Erica (December 2019). "Multiplication hits the speed limit". Communications of the ACM. 63 (1): 11–13. doi:10.1145/3371387. S2CID 209450552.
  8. ^ Cormen, Thomas H.; Leiserson, Charles E.; Rivest, Ronald L.; Stein, Clifford (2009) [1990]. "8.1 Lower bounds for sorting". Introduction to Algorithms (3rd ed.). MIT Press and McGraw-Hill. pp. 191–193. ISBN 0-262-03384-4.
  9. ^ a b c d Fredman, Michael L. (1976). "How good is the information theory bound in sorting?". Theoretical Computer Science. 1 (4): 355–361. doi:10.1016/0304-3975(76)90078-5. MR 0416100.
  10. ^ Sloane, N. J. A. (ed.). "Sequence A343245". The On-Line Encyclopedia of Integer Sequences. OEIS Foundation.
  11. ^ Lambert, Jean-Luc (1992). "Sorting the sums (xi + yj) in O(n2) comparisons". Theoretical Computer Science. 103 (1): 137–141. doi:10.1016/0304-3975(92)90089-X. MR 1181041.
  12. ^ Dietzfelbinger, Martin (1989). "Lower bounds for sorting of sums". Theoretical Computer Science. 66 (2): 137–155. doi:10.1016/0304-3975(89)90132-1. MR 1019082.
  13. ^ Kane, Daniel M.; Lovett, Shachar; Moran, Shay (2019). "Near-optimal linear decision trees for k-sum and related problems". Journal of the ACM. 66 (3): A16:1–A16:18. arXiv:1705.01720. doi:10.1145/3285953. MR 3941341. S2CID 145914158.
  14. ^ Hernández Barrera, Antonio (1996). "Finding an o(n2 log n) algorithm is sometimes hard" (PDF). Proceedings of the 8th Canadian Conference on Computational Geometry (CCCG'96). pp. 289–294.
  NODES
Idea 1
idea 1
INTERN 2
Note 1
Project 1