Because Covid, the Infinite House of Pancakes seems closed for business this year, so instead of pancake sorting we get reversorting, which is basically the same thing from another angle.
So how does Reversort work? Allow me to paraphrase the pseudocode:
Reversort(L):for i := 1 to length(L) - 1
j := argmin L[i..length(L)] Reverse(L[i..j])
How artificially iterative. Let’s reword for recursion:
As long as the section of the list under scrutiny is not known to be sorted, locate the minimum element and bring it to the front by reversal; then sort the rest of the list.
“Known to be sorted” is code for “of length 1 or less”. It’s really important to note that this specific reversal is never performed. The algorithm would work all the same if we did, and with an simpler definition too. But no.
It’s trivial to prove this it is indeed a sorting algorithm. It’s unstable, but that doesn’t matter because we’ll only ever be handling permutations of [1..N].
From this algorithm we define an algorithm cost—the number of items that undergo reversal (this is why it was important to note we’re not reversing 1-lists)—and two problems for the qualifier round:
- problem A performs the algorithm: it gives us a list and asks for the cost.
- problem C reverses1 the algorithm: it gives us a cost and asks for the list.
What order of magnitude can that cost value take?
The smallest it can be is when all reversals are as small as
possible: one element. This happens whenever the smallest sought element
is first of its sublist. In other words2, if
the list is already sorted. The total cost in that case is
The largest it can be, since the number of iterations is
predetermined, is when all reversals are as large as possible. On the
first iteration, that’s
How is this useful? Well, the problem statement gives a handful set of constraints: there are 100 tests or fewer, and the lists are 100 elements or shorter.
So assuming a linear reversal, solving problem A using a
straightforward transcription of the algorithm would perform in
The marginally tricky3 part of implementing it in Haskell is to locate the sublist minimum in a single pass. I did it like this:
-- | Extract a (minimum,sublist before,sublist after) triplet from a list.
minSplit :: [Int] -> Maybe (Int,[Int],[Int])
= Nothing
minSplit [] :t)
minSplit (x| Just (m,h',t') <- minSplit t, m < x = Just (m,x:h',t')
| otherwise = Just (x,[],t)
In plain English: an empty list has no minimum; else the result depends on how the list’s head compares to the list’s tail’s minimum. If the tail’s minimum “wins”, the result triplet is that minimum and the same post-minimum sublist, and the current item prepended to the sublist’s pre-minimum sublist. If the list’s head is a better minimum, we return it with an empty known prefix and the tail as a suffix.
Now we can implement the algorithm almost as specified:
-- | Reversort a list of ints.
reversort :: [Int] -> Writer (Sum Int) [Int]
= case minSplit xs of
reversort xs Just (m,h,t) -> do
Sum (1 + length h))
tell (:) <$> reversort (reverse h ++ t)
(m Nothing -> pure xs
I’m instrumenting with a Writer
monad to count the
reversal lengths without disrupting the algorithm flow. So problem A can
be solved with a simple getSum . execWriter . reversort
invocation.
Now for problem C. We’re given a cost, and are tasked with generating
a permutation of [1..N]
that would match costs when
reversorted. I’ll call that a (controlled) revershuffle operation. It
takes a sorted list and shuffles it, striving to consume an exact
reversal budget.
Revershuffle(N,C) returns L:1..N]
let L = [for i := N-1 downto 1
{ invariant: L[i] = N }
let P = some_amount_to_be_determined-1]) Reverse(L[i..i+P
Each reversal is intended to have cost P.4 So the total cost is going to be the sum of each step’s costs. Can we pick a distribution of them that sums to C while remaining compatible with the array bounds?
At each step, P has to be 1 or more, since that was the case for
reversort: all reversals necessarily reversed at least one element. At
step
As discussed above, a list of N elements necessarily reverses
Transcribing that back to P constraints, we’ll have
The remaining problem is merely to consume enough of the remaining budget before reaching the end. That’s a rather easy problem to solve: we can simply greedily consume as much as is available per step, until the algorithm completes. This will make for an easy check of success: the remaining budget has fallen to zero. We won’t even need to ensure our earlier computation of maximum cost is correct.
-- | Shuffle [1..n] for an exact cost of c, if possible.
revershuffle :: Int -> Int -> Maybe [Int]
= go (n - 1) [n] (c - n + 1) where
revershuffle n c | b < 0 = mzero
go _ _ b 0 l b = guard (b == 0) *> pure l
go =
go i l b let p = min (n - i) b
= splitAt p l
(before,after) in go (i - 1) (reverse (i : before) ++ after) (b - p)
It’s expressed recursively, so it reads a bit different to the
pseudocode, but it really works the same. The go
helper
takes and maintains the same i and L parameters, and an additional B one
for the remaining budget. The recursion starts with
This concludes the qualifier’s problems A and C solutions. The complete code with I/O handling is on my github.
See you soon for problem B!
Post-scriptum: the round editorial mentions that
Reversort Engineering involved an insight and working out some math.
I beg to differ. I did the bounds analysis and agree it qualifies as math, but implementing the common sense algorithm doesn’t. The upper bound doesn’t need to be pre-checked for feasibility: it’s enough to iterate, pass it, and recognize the failure.
Hah!↩︎
That word is: “inductively”.↩︎
Arguably overengineered. In a way, it’s not tricky enough: using simple linked lists as I did, there’s no way to reverse the list in the same pass as I’m looking for the minimum. In another view, it’s too much refinement when it could be simply split in more phases with the same big-O complexity: locate minimum, extract up to minimum, reverse and concatenate. The balance I struck aims at exposing as much of the original algorithm as possible. The waste here is consing or thunking the entire list in addition to the reversal, depending on what the optimizer decides.↩︎
That’s P for “price”.↩︎
It can actually only happen on the first call, but is more useful in the inner function when still debugging.↩︎