Various sorting algorithms implementations in JavaScript
Sorting Series, which is also a kind of discrete optimization problem (eg the permutation function perm
of 0..N-1
which maximizes 0*a[perm[0]]+1*a[perm[1]]+..+(N-1)*a[perm[N-1]]
is the permutation which sorts the array a
in ascending order that is a[perm[0]] <= a[perm[1]] <= .. <= a[perm[N-1]]
); lies at the center of Computer Science and Algorithms because of its many uses.
(Ref. Sorting Algorithm)
Furthermore Sorting, in one way or another, is integral part of many other important algorithms and applications (see eg. Knuth The Art Of Computer Programming)
For example Sorting is very closely associated to Searching, another topic of immense importance and applications.
Under certain sorting states, searching can be achieved in O(logN)
time or even in O(1)
time (constant) for almost every search term.
Sorting has 3 approaches:
(eg. NIST.gov maintains a dictionary of various algorithms at: http://xlinux.nist.gov/dads// )
-
In the Block case, the whole array is available at once for this case many algorithms are known (comparison-based =>
O(N^2)
,O(NlogN)
complexities) and (number/count based =>O(N)
complexity) (see below) -
In the Adaptive/Online case, the input series is accesed one at a time (for example an time-input signal). In this case some of the previous algorithms can be transformed to work adaptively
Apart from that, there are algorithms (like Dynamic Lists, Dynamic Heaps and Balanced Trees, Tries, eg AVL Trees) which keep an input sequence always in a 'sorted' state (with each new input) with relatively low complexity (eg
O(logN)
).
-
Comparison-based sorting algorithms (InsertionSort, MergeSort, QuickSort, etc..) sort a series by comparing elements with each other in some optimum sense
The best time complexity of these algorithms is
O(NlogN)
.
However better than this can be achieved -
Statistics-based sorting algorithms (CountingSort, BucketSort, RadixSort, etc..), do not use comparisons (of any kind) between elements, but instead use their arithmetic/statistical properties
This makes possible algorithms which can sort in linear
O(N)
time (the fastest possible).
However these algorithms have some limitations (eg only Integers, or special kinds of Numbers).
Is
O(N)
sorting possible for arbitrary random numbers??
Computing the value of a certain number n
(in a fixed type of encoding, eg decimal-place
) requires approximately O(logn)
"primitive digit" operations. Since, statistically, the range of values of numbers in a list of given size is increasingly correlated to the size of the list as the size of the list increases (i.e lists of size N
containing random numbers with values over the whole range 0..N
have increasingly greater probability, as N
increases, over all lists of same size N
, containing random numbers whose values do not cover the whole range 0..N
), one then has an overall complexity of O(NlogN)
even for arithmetic-based sorting algorithms.
See for example "what is the true complexity of radix sort?".
Classical algorithms for integer sorting require assumptions about the size of the integers to be sorted, or else have a running time dependent on the size.
-- Sorting in Linear Time? Arne Andersson, Torben Hagerupt, Stefan Nilsson, Rajeev Ramam
However the catch here is that same holds for comparing arbitrary numbers, computationaly one has to compare primitive digit
by primitive digit
in sequence on average, hence an additional O(logn)
complexity for comparison-based algorithms, over the O(NlogN)
bound.
Is
O(NlogN)
complexity a kind of strict base line for this computational model??
According to Knuth's theoretical lower bound theorem for general (comparison) sorting algorithms (note O(logN!) = O(NlogN)
): the O(NlogN)
bound is asymptoticaly tight (see also information-theoretic lower bound for comparison sorts is Ω(NlogN) ).
A summary of various sorting and searching algorithms can be found in SORTING AND SEARCHING ALGORITHMS, Tom Niemann (pdf)
Included Algorithms
- Builtin (JavaScript's default sorting algorithm)
- Bubble Sort
- Cocktail Sort
- Cycle Sort
- Heap Sort
- Insertion Sort
- Library Sort
- Shell Sort
- Quick Sort (recursive & iterative)
- Tree Sort
- Merge Sort (recursive, iterative & natural TODO: in-place
O(1)
extra space version) - Counting Sort
- Bucket Sort
- Radix Sort (not implemented yet)
- Intro Sort (not implemented yet)
- Burst Sort (not implemented yet)
- Tim Sort (not implemented yet)
- Permutation Sort (custom)
- Index Sort (custom)
- Statistical Sort (custom, in progress)
NOTE: The calculation of asymptotic complexity is done usually (using recursive relations) with the Master Theorem :
Refs. http://en.wikipedia.org/wiki/Master_theorem, http://en.wikipedia.org/wiki/Introduction_to_Algorithms
T(n) = aT(n/b) + f(n), a>=1, b>1
eg. for MergeSort => T(n) = 2T(n/2) + O(n) => T(n) = O(nlogn)
This package implements showcases of various (best known) sorting algorithms (and a couple of custom ones) for study, experimentation and use in applications In a concice library
Algorithms as a technology Suppose computers were infinitely fast and memory was free. Would you have any reason to study algorithms? The answer is yes, if for no other reason than that you would still like to demonstrate that your solution method terminates and does so with the correct answer. ...Of course, computers may be fast but not infinitely fast and memory may be cheap but not completely free. Computing time is therefore a bounded resource, and so is space in memory. These resources should be used wisely and algorithms that are efficient in terms of time and space will help you do so. This demostrates that algorithms, like computer hardware, are a technology . Total system performance depends on choosing efficient algorithms as much as choosing fast hardware. Just as rapid advances are being made in other computer technologies, they are being made in algorithms as well. (Introduction to algorithms, 2nd Ed. Cormen,Leiserson,Rivest,Stein)
Algorithms as a ecological technology Additionaly, every operation/instruction a computer performs has an energy consumption cost. Thus an efficient algorithm saves energy! An efficient algorithm performs a computation by trying to use the resources in the best possible manner, so effectively uses energy in the best possible manner. Where does energy come from? It comes from burning coal or gas (mainly). So there you have it; efficient code is ecological! Better start learning your complexity soon.
see also:
- Abacus advanced Combinatorics and Algebraic Number Theory Symbolic Computation library for JavaScript, Python
- Plot.js simple and small library which can plot graphs of functions and various simple charts and can render to Canvas, SVG and plain HTML
- HAAR.js image feature detection based on Haar Cascades in JavaScript (Viola-Jones-Lienhart et al Algorithm)
- HAARPHP image feature detection based on Haar Cascades in PHP (Viola-Jones-Lienhart et al Algorithm)
- FILTER.js video and image processing and computer vision Library in pure JavaScript (browser and node)
- Xpresion a simple and flexible eXpression parser engine (with custom functions and variables support), based on GrammarTemplate, for PHP, JavaScript, Python
- Regex Analyzer/Composer Regular Expression Analyzer and Composer for PHP, JavaScript, Python
- GrammarTemplate grammar-based templating for PHP, JavaScript, Python
- codemirror-grammar transform a formal grammar in JSON format into a syntax-highlight parser for CodeMirror editor
- ace-grammar transform a formal grammar in JSON format into a syntax-highlight parser for ACE editor
- prism-grammar transform a formal grammar in JSON format into a syntax-highlighter for Prism code highlighter
- highlightjs-grammar transform a formal grammar in JSON format into a syntax-highlight mode for Highlight.js code highlighter
- syntaxhighlighter-grammar transform a formal grammar in JSON format to a highlight brush for SyntaxHighlighter code highlighter
- SortingAlgorithms implementations of Sorting Algorithms in JavaScript
- PatternMatchingAlgorithms implementations of Pattern Matching Algorithms in JavaScript