(This is a beta version of editorial. It will be changed later)

I hope everybody enjoyed the contest. Here we'll present some of the author's solutions to the problems together with some other approaches that we came up with. The editorial is prepared by Lasha Lakirbaia, Anton Lomonos, and Ievgen Soboliev.

# A — Game

(developed by Ievgen Soboliev & Eldar Bogdanov, editorial written by Lasha Lakirbaia)

The problem can be solved with dynamic programming technique. Let *winner*[*a*][*b*][*move*] denote who wins if both players play optimally and the first has a balls left in his box, the second has *b* balls left in his box and move defines who is to move next. To determine the value of *winner*[*a*][*b*][*move*], we should consider all the possible moves the next player can make and see whether any of them lead into the position *winner*[*x*][*y*][*move*2] in which this player can win with optimal play. Thus, values of winner can be defined recursively.

Ok it was a joke, the solution is just this one line: **println (n1 > n2 ? “First” : “Second”)**.

# B — Permutations

(developed by Ievgen Soboliev & Eldar Bogdanov, editorial written by Lasha Lakirbaia)

For a given array *p* _{1}, *p* _{2}, ..., *p* _{ k} let’s call the value of that expression the *value* of that array.

For the first subproblem, we can just consider all *n*! permutations in lexicographic order and keep 2 variables — one of them should be the highest value so far, and the other should be the list of permutations with that highest value. Than we can update this variables like this: When he have a new permutation, if it has smaller value, we disregard it. If it has larger value, we assign the first variable this new value and to the second variable we assign 1-element list, containing this current permutation only. In the end we just pick *m*-th element from the second variable. From now on we’ll concentrate on solving the harder one.

First of all let’s find out which permutations have the highest values among all the permutations of (1, 2, ..., *n*). It's easy to note that whenever we have some *n*-element array, in the highest valued permutations the position of the smallest element of the array will be either first or the last in the permutation. This can be proved easily as follows: There can't be an optimal (highest valued) permutation where the smallest element is in the middle (I mean after the first and before the last), because moving this element to either beginning or ending would obviously increase the value, that would thus contradict the statement that the original permutation had the highest value. So we proved that the position of 1 should either be at the beginning or at the end of a highest valued permutation. Moreover, it’s easy to see that in both cases, the value will be the same. After we decide on which end 1 will be, than we can move on to the remaining array (2, 3, ..., *n*) and do the same for 2 as we did for 1. Than we go on and do the same for 3, 4, ..., *n*.

From here we can clearly say what are all the permutations with the highest value — all the permutations which can be derived by the following algorithm: Consider numbers (1, 2, ..., *n*) in the reverse order. Each time add the current number to one of the ends of the current permutation (initially the permutation is empty). For example, for n=5 the possible permutations are (1, 2, 3, 4, 5), (1, 3, 5, 4, 2), (5, 4, 3, 2, 1) and (1, 2, 5, 4, 3). One more important thing to note is that the number of highest valued permutations of any n-element array is equal to 2^{ n - 1}, which easily follows from the way we’re constructing all these permutations. With all that said, the rest of the solution is pretty straightforward. Let’s consider numbers 1, 2, ..., *n* in this order and apply the following recursive approach: If we’re currently on **pos**-th position and searching for *m*-th element, having already decided the indices of all the elements less than pos, then there are 2 cases: We know that number of possible permutations is 2^{ n - pos - 1}. If *m* ≤ 2^{ n - pos - 1}, then we put pos element in the beginning of the array and recursively go to the state (*pos* + 1, *m*). Otherwise, we’ll put pos in the end of the array and recursively go to the state (*pos* + 1, 2^{ n - pos - 1} - *m*). The total running time of this solution is *O*(*n*), if implemented in optimal way but almost any kind of implementation of this idea would work, as the value of *n* is relatively small.

*Note 1: From the solution it's visible that m ≤ 2^{49}, and also for n = 50 m can be as big as 2^{49}, which doesn't fit in int but fits in long.*

*Note 2: This last steps can also be implemented by just considering the binary representation of m.*

# C — Second Price Auction

(developed by Ievgen Soboliev & Eldar Bogdanov, editorial written by Lasha Lakirbaia)

This problem can be solved in many different ways. Here we’ll explain 3 of them. Let’s denote by *P*(*b*) the probability of second highest bid being equal to *b*. Obviously, possible values of *b* are only integers from 1 to 10, 000. If we can calculate *P*(*b*) for each *b* in that range, then the desired expected value is just . Now let’s concentrate on calculating *P*(*b*) for a given *b*. These 3 solutions have different approaches to calculating the value of *P*(*b*).

First solution (the easiest) — brute force

Let’s consider every partition of a set of bidders into 3 subsets, such that in the first set each bidder had higher bid than *b*, in second set each bid was equal to *b*, while in the third set each bid was less than *b*. As we assume that *b* is the second highest bid, it’s easy to notice that the first set contains either 0 or 1 element, while the second set contains at least 1 element. (Also, if the first set is empty, the second set must contain at least 2 elements) Number of such partitions is *O*(2^{ n}·*n*). After we have this partition, for each bidder we know if its bid should be more, equal or less then *b*. Probability of any of these can easily be counted in *O*(1). So in the end we get the complexity of *O*(2^{ n}·*n*·*max*(*R* - *L* + 1)), which is good enough for given constraints.

Second solution — dynamic programming

Let’s consider the following dynamic programming approach. Denote by **(pos, moreCount, equalCount)** the state, when we already considered all bidders with indices less then **pos**, the second largest bid in the end must be equal to *b*, currently the number of bidders with bid higher than *b* is **moreCount** and the number of bidders with bid equal to *b* is **equalCount**. Finally, *dp*(*pos*, *moreCount*, *equalCount*) denotes the value of *P*(*b*) for this state. Note, that we don’t have to consider the states where *moreCount* > 1, because for them *P*(*b*) = 0. Also, we don’t need to distinguish between states where some of them has *equalCount* = 2 and some of them has *equalCount* > 2. They have same values. So we’ll assume that *equalCount* < 3 and *moreCount* < 2.

Transitions for this dynamic programming are as follows: - Consider current bid to be higher than *b*. It’s only possible if *moreCount* = 0, because otherwise the probability of *P*(*b*) will be 0. If *moreCount* = 0, the transition will be to *dp*(*pos* + 1, 1, *equalCount*). - Consider current bid to be equal to *b*. It will lead us to a transition to *dp*(*pos* + 1, *moreCount*, *min*(*equalCount* + 1, 2)). - Consider current bid to be less than *b*. Transition in the case will be *dp*(*pos* + 1, *moreCount*, *equalCount*).

All these cases have their respective probabilities, which are straightforward to calculate in *O*(1). So as we see, the number of states of dynamic programming is O(n), while calculating each state take *O*(1). So in total, calculating *P*(*b*) takes *O*(*n*) time, which leads us to the solution of the initial problem in *O*(*n* * *max*(*R* - *L* + 1)).

Third solution — pure math

*P*(*b*) = *Pr*.(*maxBid* = *b*, *secondMaxBid* = *b*) + *Pr*.(*maxBid* > *b*, *secondMaxBid* = *b*). Let’s calculate these 2 summands separately. Denote by *r* _{ i} a random bid of *i* - *th* bidder.

*Pr*.(*maxBid* = *b*, *secondMaxBid* = *b*) = *Pr*.(*maxBidb*) - *Pr*.(*maxBid* < *b*) . Now let’s calculate the other summand: .

Note that each of *Pr*.(*r* _{ i} > *b*), *Pr*.(*r* _{ i} = *b*) and *Pr*.(*r* _{ i} < *b*) can be calculated in *O*(1), so *P*(*b*) can be calculated in *O*(*n* ^{2}). So in the end, we get the running time of *O*(*n* ^{2}·*max*(*R* - *L* + 1)). However, the last step can easily be implemented in *O*(*n*), which would give us the total running time of *O*(*n*·*max*(*R* - *L* + 1)), but in this problem it’s not required as *n* ≤ 5.

Note: As you can see, using second and third solutions the problem could be solvable for much larger values of *n*. However, we felt like all three approaches are equally interesting and let any of them pass. Bonus: assuming that we can multiply doubles without precision loss in polynomial time, can you solve the problem in polynomial time if the bid ranges of each company are unbounded?

# D — Constrained Tree

(developed by Eldar Bogdanov, Alexander Ruff and Anton Lomonos, editorial written by Anton Lomonos)

Let's try to construct this tree from top to bottom trying to comply with given set of constraints. The first observation you need to make is if vertices *i* and *j* (*i* < *j*) should be in a given subtree then any vertex *k* (*i* < *k* < *j*) should also be in this subtree. It leads to 2 important conclusions:

- Our task "find a tree rooted at vertex 1 which should contain
*N*vertices and comply to the given set of constraints" can be rephrased to "find a tree rooted at vertex 1 which should comply to the given set of constraints, contain the least number of vertices as possible, and contain vertex*N*". - We need to know only minimums and maximums for left and right parts of each subtree.

Let's assume we are trying to construct subtree rooted at *i* that should contain vertex *m*, vertices [*j* _{1}..*j* _{2}] to the left of the root, and vertices [*k* _{1}..*k* _{2}] to the right of the root. If we don't have constraints for the left part or for the right part of the subtree then we can place all the vertices utterly to the right or to the left (we still need to check constraints for sanity). Otherwise, the conditions *i* < *j* _{1}, *j* _{2} < *k* _{1} should take place. The vertex *i* + 1 should be the left child of the root. Vertices [*j* _{2} + 1;*k* _{1} - 1] may be placed to the left or to the right which depends on the constraints. Let's construct the left subtree rooted at *i* + 1 in such a way that it should contain the vertex *j* _{2} and contain at least number of vertices as possible. Let's assume the vertex with the biggest label in this subtree is *mid*. Condition *mid* < *k* _{1} should take place. Then we can make right child of the root to be *mid* + 1, and we need to construct subtree rooted at *mid* + 1 that should contain at least number of vertices as possible and should contain vertex *max*(*k* _{2}, *m*)

The complexity of algorithm is *O*(*N* + *C*).

# E — Subarray Cuts

(proposed by Nikhil Goyal, developed by Anton Lomonos and Lasha Lakirbaia, editorial written by Anton Lomonos and Lasha Lakirbaia)

This problem can be solved in several ways. One of the expected solution for the easier part has complexity *O*(*n* ^{2} *k*), which will be presented later. For now let's discuss another solution, which is also good enough for first subproblem and has complexity of *O*(*nk* ^{2}).

Let’s call the set of all possible *k*-tuples of partitions *A*; So we’re concentrating on finding the element of *A* with the maximal sum value. Let’s consider another set of tuples of partitions but with the following properties: We have at most *k* disjoint subarrays chosen. Call their sums *s* _{1}, *s* _{2}, …, *s* _{ m} ( *m* is the number of subarrays and *m* ≤ *k*). For each 1 ≤ *i* ≤ (*m* - 1), (*s* _{}(*i* - 1) - *s* _{ i}) * (*s* _{ i} - *s* _{}(*i* + 1)) < 0. In other words, *s* _{ i} is **not between** its neighbors by value. (For example, if its neighbors are 3 and 11, it will be either greater than 11 or less than 3). The number of unused (not picked) elements from the initial array is at least (*k* - *m*). In other words, we are able to add enough disjoint subarrays to this *m* subarrays so that in the end we have exactly *k* of them. Let’s call this set *B*.

It can be easily proved, that the maximal valued elements in *A* and *B* have the same value. The proof of this statement is given in the very end of the solution. For now let’s believe it and solve the problem.

So from now on we concentrate on finding the maximal value with the property of *B*.

Note that the value of any element of B with sums *s* _{1}, *s* _{2}, …, *s* _{ m} is equal to the maximum of ( + *s* _{1} - 2·*s* _{2} + 2·*s* _{3} - 2·*s* _{4} + …) and ( - *s* _{1} + 2 *cdots* _{2} - 2 *cdots* _{3} + 2·*s* _{4} - …). The important catch here is that we can ignore the property of *B* which says that each sum of subarray is NOT between it’s neighboring sums, because in the end we’ll get the same answers. So we just ignore that property and solve the problem without that restriction. The only restriction remaining is that they should be disjoint and the total number of unused elements should be at least (*k* - *m*). Here is the sketch of the proof why will it give the same answer: *s* _{1} - 2·*s* _{2} + 2·*s* _{3} - … = (*s* _{1} - *s* _{2}) + (*s* _{3} - *s* _{2}) + … ≤ |*s* _{1} - *s* _{2}| + |*s* _{2} - *s* _{3}| + …, as well as ( - *s* _{1} + 2·*s* _{2} - 2·*s* _{3} + ...). It means that when they’re both maximum, they equal to each other.

So we can instead concentrate on finding the maximal possible ( + *s* _{1} - 2·*s* _{2} + 2·*s* _{3} - 2·*s* _{4} + …) and ( - *s* _{1} + 2·*s* _{2} - 2·*s* _{3} + 2·*s* _{4} - …) values and it will coincide with the answer to our problem.

Denote by *dp*(*pos*, *sign*, *first*, *last*, *unused*, *max*) — the answer for the subproblem (*a* _{ pos}, *a* _{}(*pos* + 1), …) where *sign* means the sign of the first chosen subarray, *first* and *last* denote whether the current subarray should be first, last or none of them, *unused* denotes the number of unused elements that we should preserve and *max* denotes the maximum number of subarrays we may take. There are *O*(*nk* ^{2}) states here, cause *sign*, *first* and *last* each have just 2 options. Important part here is how we make transitions in *O*(1) and thus solve the entire problem in *O*(*nk* ^{2}). First transition would correspond to ignoring *pos*-th element — *dp*(*pos* + 1, *sign*, *first*, *last*, *min*(*unused* + 1, *k*), *max*). Otherwise we are taking *a* _{ pos} in the first chosen subarray. There are 2 cases here: Either the current interval is just (*pos*, *pos*), or it’s length is more than 1. In the first case, the value is *signa* _{ pos}·((*firstorlast*)?1: 2) + *dp*(*pos* + 1, - *sign*, *false*, 0*or*1, *unused*, *max* - 1). In the second case, the value would be *sign*·*a* _{ pos}·((*firstorlast*)?1: 2) + *dp*(*pos* + 1, *sign*, *first*, *last*, *unused*, *max*).

Hence, in total we have 3 transitions and thus we calculate each *dp*(*pos*, *sign*, *first*, *last*, *unused*, *max*) in *O*(1). It means the entire complexity of the solution is *O*(*nk* ^{2}). Note: We didn’t present every detail and corner case, because it would be very boring to read otherwise.

Below we’ll present the solution for the harder version of this problem.

In general case we are trying to calculate this expression:

*max*(... + |*s* _{ j - 2} - *s* _{ j - 1}| + |*s* _{ j - 1} - *s* _{ j}| + |*s* _{ j} - *s* _{ j + 1}| + ...) = *max*(... ± (*s* _{ j - 2} - *s* _{ j - 1}) ± (*s* _{ j - 1} - *s* _{ j}) ± (*s* _{ j} - *s* _{ j + 1}) ± ...)

This task can be solved by using dynamic programming. Let's try to calculate recurrently the following function where we should pick subarrays *j*..*k* from array [*i*; *n*]:

*f*(*i*, *j*, *c* _{1}, *c* _{2}) = *max*(... ± (0 - 0) *c* _{1} (0 - *s* _{ j}) *c* _{2} (*s* _{ j} - *s* _{ j + 1}) ± (*s* _{ j + 1} - *s* _{ j + 2}) ± ...), where are defined.

The recurrent formula for this function is:

*f*(*i*, *j*, *c* _{1}, *c* _{2}) = *max* _{ p 1, p 2}(*c* _{1} ( - *sum* _{ p = p 1..p 2} *a* _{ p}) *c* _{2} (*sum* _{ p = p 1..p 2} *a* _{ p}) + *f*(*p* _{2} + 1, *j* + 1, *c* _{2}, ± ))

We can easily get rid of using *p* _{1}:

*f*(*i*, *j*, *c* _{1}, *c* _{2}) = *max*(*f*(*i* + 1, *j*, *c* _{1}, *c* _{2}), *max* _{ p} _{2}(*c* _{1} ( - *sum* _{ p = i..p 2} *a* _{ p}) *c* _{2} (*sum* _{ p = i..p 2} *a* _{ p}) + *f*(*p* _{2} + 1, *j* + 1, *c* _{2}, ± ))

We can get rid from *p* _{2} by using additional function *g*(*i*, *j*, *c* _{1}, *c* _{2}), which states that subarray *j* already has some elements in it:

*f*(*i*, *j*, *c* _{1}, *c* _{2}) = *max*(*f*(*i* + 1, *j*, *c* _{1}, *c* _{2}), *c* _{1} ( - *a* _{ i}) *c* _{2} *a* _{ i} + *g*(*i* + 1, *j*, *c* _{1}, *c* _{2}))

*g*(*i*, *j*, *c* _{1}, *c* _{2}) = *max*(*f*(*i*, *j* + 1, *c* _{2}, + ), *f*(*i*, *j* + 1, *c* _{2}, - ), *c* _{1} ( - *a* _{ i}) *c* _{2} (*a* _{ i}) + *g*(*i* + 1, *j*, *c* _{1}, *c* _{2}))

There is still a matter of handling the beginning and ending of dynamic programming. There are several ways to do it. For example, we can calculate maximums and minimums of values *s* _{1} or *s* _{ k} based on their endings or beginnings, or we can make exceptions for using *c* _{1} and *c* _{2} if *j* = 1 or *j* = *k*.

The complexity of the algorithm is *O*(*n* * *k*).

# F — Scaygerboss

(developed by Anton Lomonos, editorial written by Anton Lomonos)

The task reduces to the situation when we have *k* females and *k* males, and we need to connect them in such a way, that there will be no more than one pair in each cell. This task can be solved by using maximum flow algorithm.

Firstly we need to choose the value of our answer *ans* (for example, by binary search) and check if they can connect in a time no more than *ans*. Let's make edges of capacity 1 between source and females, between males and sink, and between inputs and outputs of each cell. Then let's make edges of capacity 1 between females and inputs of cells and between outputs of cells and males if they can reach this cell within *ans* (we can calculate these values by using breadth-first algorithm). If we can get a maximum flow of capacity *k* on this graph then they can reach each other within *ans*.

The above approach is enough to solve the first part of the task. It has complexity of *O*(*k* ^{3}·*log*(*k*)). For solving second part of the task one should optimize the above solution. Highly optimized solution with complexity of *O*(*k* ^{3}·*log*(*k*)) could pass, but there exists a solution with complexity *O*(*k* ^{3}).

The first observation you need to make is you don't need to reconstruct the graph each time for running maximum flow algorithm. If you run the algorithm for *ans* = *a*, and then run the algorithm for *ans* = *b*, *b* > *a*, you would just need to add additional edges and "continue to run" maximum flow algorithm. The complexity of this would be *O*(*k* ^{3} + *l*·*k* ^{2}) where *l* is the number of times you do this trick. Using this observations in different ways could be enough to pass the second part of the task.

You can achieve the complexity of *O*(*k* ^{3}) by using square root decomposition. You need to divide *k* ^{2} edges to *k* batches of *k* edges based on the values of *ans* you can use these edges for. Running maximum flow algorithm after adding each of these batches would get you *O*(*k* ^{3}) complexity. Then you know which batch would tell you what the answer is, and you run maximum flow algorithm for this batch, which is *O*(*k* ^{3}) complexity too.

You can also divide edges to batches not based by their quantities, but based by their values. You will be adding the edges of the same values to the graph in the same blow. That way your algorithm would run even faster.

I (White_Bear) would like to give special congratulations to mmaxio, Petr, niyaznigmatul, and Egor who solved F2 using Java.

# G — Inversions Problem

(proposed by Ievgen Soboiliev, developed by Ievgen Soboiliev, editorial written by Lasha Lakirbaia, to be changed)

This is a draft of solution to this problem. It will be changed later. Currently it contains solutions of first 2 subproblems.

The first subproblem can be solved many ways. The easiest probably is the following: consider the dynamic programming approach where the state is (*perm*, *m*) — where *perm* is the current permutation and *m* is the number of operations that are going to be performed over it. Transitions are just all the possible permutations that can be derived in the next turns (with their respective probabilities) and obviously the other variable will be just (*m* - 1). The running time for this approach is *O*(*n*! * *n* ^{3} * *k*), which is good enough for the given constraints. Note that even though for small *n* the number of possible permutations is pretty small and *k* is also small, solutions based on simulation are not working at the precision level that’s required. To solve 2 other subproblems, we are gonna need more observations. First one is the linearity of expected value. It basically means that the expected value of a sum of random variables is equal to the sum of expected values of those variables, even if they’re somehow dependent on each other. Let’s see how we can use this idea to solve this problem. For this we have to look at the number of inversions a bit differently. For each possible pair (*i*, *j*) with 1 ≤ *i* < *j* ≤ *n*, let’s define the random variable *b*(*i*, *j*) to be 0 if *i* - *th* and *j* - *th* elements will preserve the order after *k* operations and 1 otherwise. Now, the number of inversions in terms of these variables is equal to . So if we can find the expected values of each *b*(*i*, *j*) variables, than calculating the answer will be straightforward.

Another observation is that the expected value of a binary variables is the same as it’s probability of being equal to 1. It means that the expected value of *b*(*i*, *j*) is same as probability that *i* - *th* and *j* - *th* elements will swap the order in the end, after *k* moves.

From now on, let’s denote by *p*(*i*, *j*, *m*) the probability, that after m steps the *i* - *th* and *j* - *th* elements will swap the order (in other words, probability that the final position of the *i* - *th* element will be to the right side of the final position of *j* - *th* element). The solutions of second and third subproblems differ only in ways of calculating these values.

Easier way to find these values is to use recursive approach: Consider all the possible next moves — namely all possible intervals (*x*, *y*) on which the next step will be performed. Each of them has probability equal to . There are several cases:

(1) Intervals (*x*, *y*) and (*i*, *j*) are either disjoint, or *i* < *x*, *y* < *j*. In both cases, the transition will be just *p*(*i*, *j*, *m* - 1).

(2) *x* ≤ *i* ≤ *y* < *j*. In this case the transition is gonna be *p*(*x* + *y* - *i*, *j*, *m* - 1).

(3) *i* < *x* ≤ *j* ≤ *y*. In this case the transition will be *p*(*i*, *x* + *y* - *j*, *m* - 1).

(4) *x* ≤ *i* ≤ *j* ≤ *y*. In this case the transition will be *p*(*x* + *y* - *i*, *x* + *y* - *j*, *m* - 1).

The running time of this solution will be *O*(*n* ^{4} * *k*), which is good enough for 2-nd subproblem, but not for the third one. This solution can actually be optimized. For this, we will need to deal with above 4 cases in *O*(*n*) each, instead of *O*(*n* ^{2}).

(1) We just count how many (*x*, *y*) intervals are of this type (*i* < *x*, *y* < *j*) and than multiply this number by *p*(*i*, *j*, *m* - 1).

(2) For each fixed possible value of (*x* + *y*), count how many (*x*, *y*) exist with that sum such that *x* ≤ *i* ≤ *y* < *j*. It can easily be done in *O*(1). As there are *O*(*n*) possible values of (*x* + *y*), this whole step takes O(n).

(3) The approach here is exactly the same as in the second case.

(4) For each possible value of *s* = *x* + *y* such that *x* ≤ *i* ≤ *j* ≤ *y* and sum up (1 - *p*(*s* - *j*, *s* - *i*, *m* - 1)). There are *O*(*n*) possible values of *s*, so this whole step takes *O*(*n*) time.

As these 4 steps each take at most *O*(*n*), the total running time will be *O*(*n* ^{3}·*k*). The fastest solution works in *O*(*n* ^{2}·*min*(*k*, 2000)). This solution will be described later.

Very quick editorial

Hello =)

I have one thing to say about the first subproblem of the problem G, more specific, about: "solutions based on simulation are not working at the precision level that’s required". My solution only simulates the process, and it got an accepted veredict, maybe is that I count the inversions, and after all the process, I divide the number of inversions by the number of possible 'configurations'. (Let me know if you don't understand my english).

This is my code if you want to check it: 9762896

I think what's meant by simulation is the following (different from yours — which I would call brute force):

ohh, thanks, now it's clear :D

asdf00001 Thanks for the feedback. But yes, I_love_Hoang_Yen is correct. In the editorial I meant the solutions where you do a large number of simulations of the given process, sum up the number of inversions in each of them and then count the average result and output as an answer.

For Problem B,can someone please explain this?

`Note 2: This last steps can also be implemented by just considering the binary representation of m.`

Present m-1 as binary number contains n-1 digits. For i = 1:n-1 we do this: if i-th digit is 0, then we put i at the begin of answer, else — at the end.

Example: n = 5, m = 13. 12 = (1100)2; Answer is (? ? ? ? ?).

1-st digit of m is 1 -> Answer is (? ? ? ? 1)

2-nd digit of m is 1 -> Answer is (? ? ? 2 1)

3-rd digit of m is 0 -> Answer is (3 ? ? 2 1)

4-th digit of m is 0 -> Answer is (3 4 ? 2 1)

And 5 we put at se second place: (3 4 5 2 1). This is answer. Why it's work? I don't know. Just noticed regularity.

It works, because (in the editorial's solution) we put the number to end only if m is larger than some power of 2 (it will be the largest power of two smaller than m) and then we subtract this power. So, we see MSB 1 in binary, we subtract it from m and put the number to the end, we see 0 in binary, we put the number to the beginning.

In problem C, formula pictures are very ugly and also there are mistakes, need Product instead of Sum in some cases.

I agree that formulas are ugly (I tried to fix that without success), but tell me where do you think that there is product instead of sum?

Pr(maxBid=b, secondMaxBid=b) = ... =

PRODUCTPr(Ri <= b) —PRODUCTPr(Ri < b) — ...Prob. that

allare smaller equalminus prob. that

allare smaller (0 equal)minus prob. that one is equal and others are smaller

Oh thanks, I missed it. I'll fix them.

Can someone please explain the solution to problem-E? I can't understand the formulation of DP solution presented in the editorial.

It's a little bit complicated, but I'll try to explain further. We made the following observation first (before DP): We can just pick any

msubarrays wherem≤kand the number of unused elements are at least (k-m). (The idea is that we can potentially add (k-m) 1-length subarrays to get exactlykin the end, as the problem statement forces us). So after we have the observation, second trick is that we want to get rid of absolute values and replace them with just brackets, which will make it easier for us to solve the problem. Why is this the correct approach and why it produces correct result is explained in the editorial. If this part is also unclear, I can explain further. So to conclude: We considermsubarrays such thatm≤kand number of unused elements are at least (k-m). If we denote their sums ass_{1},s_{2}, ...,s_{ m}respectively, we will try to find the maximal possible value of the following:max(( +s_{1}- 2·s_{2}+ 2·s_{3}- 2·s_{4}+ ... + / - 2·s_{}(m- 1) - / +s_{ m}), ( -s_{1}+ 2·s_{2}- 2·s_{3}+ 2·s_{4}- ... + / - 2·s_{}(m- 1) - / +s_{ m}). (The sign of last coefficient depends on the parity ofm).Now with all that said, dp state is the following: We are currently on

pos-th position and we've already made decision of all array members with indexes less thenpos; The sign of the next picked subarray will besign; The next picked subarray will befirstorlastor none of them (we need this to determine the coefficient — 1 or 2 with respective sign); Currently the count of unused elements is equal tounused; the maximum amount of subarrays that we can still pick is equal tomax;Please try to understand the editorial about transitions for this dp. If it's still unclear, let me know :)

It's all clear now. Thanks for the reply!

Could somebody explain Problem's C first solution, why can we calculate probability for each of 2^n sets, if not all bidders take part in the auction in the current set?

I'll write in details. We consider each possible split set of all bidders into 3 (possibly empty) subsets. Call them

A,BandC. We assume that every bidder inAhas made bid strictlymorethanb, each member ofBmade a bidequaltoband each member ofChas bid strictly less thanb. Now for each split, we calculate it's probability. First of all, we are not interested in splits, whereAcontains more than 1 element, because it would mean thatbis no longer a second best. Also, setBcan't be empty. One more thing to check is that ifAis empty, thanBmustcontain at least 2 elements. If all of these constraints are satisfied, then we just calculate probability the following way: For each member ofA, calculate the probability that it's bid is more thanband multiply these probabilities, as they're mutually independent of each other. (Product of an empty set is supposed to be 1) Let's do the similar thing forBandCand in the end just multiply all these probabilities. Then we have an outer loop to sum up all those products and it's gonna be our answer.One more important thing to note is how we implement the splitting into 3 parts and how many splits are there. As we said that

Ais either empty or contains just a single element, we have (n+ 1) choices orA: Either be empty, or contain any bidder out ofnpossibilities. Once we pickA, we have either (n- 1) ornmore bidders. Than we have either 2^{ n - 1}or 2^{ n}possibility forB. After we pickAandB, the others will automatically go toC. From here we see that number of states isO(2^{ n}·n) and we also see the way to implement it easily: First we make an outer loop forA, and then inside we make inner loop of bitmasks forB, after whichCbecomes fixed.In problem B second solution,when m > 2^(n - pos - 1), we should go to state (pos + 1, 2^(n - pos - 1) - m)? ... i think we should go to state (pos + 1, m — 2^(n — pos — 1) — m) instead.

That's right, thanks for pointing out. I'll edit.

m — 2^(n — pos — 1) — m = — 2^(n — pos — 1)

sorry what i meant is

`m-2^(n-pos-1)`

I agree, It should be modified to

m- 2^{ n - pos - 1}. The tutorial is still not updated. The tutorial lets m goes to negative value in the recursive call. Check my accepted submission usingm- 2^{ n - pos - 1}at 13916702Sorry, I didn't explain what I don't understand. I mean why we can add to the answer the probability of b — second number of the current set, if other bidders, that are out of the current set, can do this b impossible.

It is hard for me to speak eng, so if you don't understand me, here example:

if we have mask =

`0011`

(bits), set B wil consist of`{(1, 5), (1, 6)}`

. So we will add to the answer wrong value, because`b`

must be`>= 7`

. So can you explain what mean the number, that we calculate for each set?If it's impossible, the probability will be 0 and we will add 0 to the answer.

I don't understand the mean of problem D; Can you show through a data to tell your algorithm?

I am a novice programmer. I am having trouble understanding the problem G. I don't get how the probability is calculated here. Could anyone give me the explanation? Thanks in advance :)

Hi,there is something I don't quite understand yet. In problem B Permutations you wrote "Otherwise(which is the condition that m > 2^(n - pos - 1)), we’ll put pos in the end of the array and recursively go to the state (pos + 1, 2^(n - pos - 1) — m)" However, shouldn't the upper bound be m — 2^(n — pos — 1) since m is larger than 2^(n - pos - 1)?

Heh, it is funny that it is said that standard solution to F works in and you described how to improve it in

O(k^{3}), but in fact Dinic on graphs with unit capacity works in , so in fact my standard Dinic without any problem-specific optimizations results in :)).In G3, why is the average number of inversions tends to a constant with increasing of k? It feels true but I cannot come with a strict mathematical proof of it.

Will there be solution of problem G3?

Where is the O(n^2·min(k, 2000)) solution for G3? Since O(n^3·k) is not good enough for G3.

Please help with this:)

The editorial for the first problem really got me.