rekt_n00b's blog

By rekt_n00b, history, 4 years ago, , Introduction

Mo's Algorithm has become pretty popular in the past few years and is now considered as a pretty standard technique in the world of Competitive Programming. This blog will describe a method to generalize Mo's algorithm to maintain information about paths between nodes in a tree.

Prerequisites

Mo's Algorithm — If you do not know this yet, read this amazing article before continuing with this blog.

Preorder Traversal or DFS Order of the Tree.

Problem 1 — Handling Subtree Queries

Consider the following problem. You will be given a rooted Tree T of N nodes where each node is associated with a value A[node]. You need to handle Q queries, each comprising one integer u. In each query you must report the number of distinct values in the subtree rooted at u. In other words, if you store all the values in the subtree rooted at u in a set, what would be the size of this set?

Constraints

1 ≤ N, Q ≤ 105

1 ≤ A[node] ≤ 109

Solution(s)

Seems pretty simple, doesn't it? One easy way to solve this is to flatten the tree into an array by doing a Preorder traversal and then implement Mo's Algorithm. Maintain a lookup table which maintains the frequency of each value in the current window. By maintaining this, the answer can be updated easily. The complexity of this algorithm would be Note that you can also solve this in by maintaining a set in each node and merging the smaller set into the larger ones.

Problem 2 — Handling Path Queries

Now let's modify Problem 1 a little. Instead of computing the number of distinct values in a subtree, compute the number of distinct values in the unique path from u to v. I recommend you to pause here and try solving the problem for a while. The constraints of this problem are the same as Problem 1.

The Issue

An important reason why Problem (1) worked beautifully was because the dfs-order traversal made it possible to represent any subtree as a contiguous range in an array. Thus the problem was reduced to "finding number of distinct values in a subarray [L, R] of A[]. Note that it is not possible to do so for path queries, as nodes which are O(N) distance apart in the tree might be O(1) distance apart in the flattened tree (represented by Array A[]). So doing a normal dfs-order would not work out.

Observation(s)

Let a node u have k children. Let us number them as v1,v2...vk. Let S(u) denote the subtree rooted at u.

Let us assume that dfs() will visit u's children in the order v1,v2...vk. Let x be any node in S(vi) and y be any node in S(vj) and let i < j. Notice that dfs(y) will be called only after dfs(x) has been completed and S(x) has been explored. Thus, before we call dfs(y), we would have entered and exited S(x). We will exploit this seemingly obvious property of dfs() to modify our existing algorithm and try to represent each query as a contiguous range in a flattened array.

Modified DFS-Order

Let us modify the dfs order as follows. For each node u, maintain the Start and End time of S(u). Let's call them ST(u) and EN(u). The only change you need to make is that you must increment the global timekeeping variable even when you finish traversing some subtree (EN(u) = ++cur). In short, we will maintain 2 values for each node u. One will denote the time when you entered S(u) and the other would denote the time when you exited S(u). Consider the tree in the picture. Given below are the ST() and EN() values of the nodes. ST(1) = 1 EN(1) = 18

ST(2) = 2 EN(2) = 11

ST(3) = 3 EN(3) = 6

ST(4) = 4 EN(4) = 5

ST(5) = 7 EN(5) = 10

ST(6) = 8 EN(6) = 9

ST(7) = 12 EN(7) = 17

ST(8) = 13 EN(8) = 14

ST(9) = 15 EN(9) = 16

A[] = {1, 2, 3, 4, 4, 3, 5, 6, 6, 5, 2, 7, 8, 8, 9, 9, 7, 1}

The Algorithm

Now that we're equipped with the necessary weapons, let's understand how to process the queries.

Let a query be (u, v). We will try to map each query to a range in the flattened array. Let ST(u) ≤ ST(v) where ST(u) denotes visit time of node u in T. Let P = LCA(u, v) denote the lowest common ancestor of nodes u and v. There are 2 possible cases:

Case 1: P = u

In this case, our query range would be [ST(u), ST(v)]. Why will this work?

Consider any node x that does not lie in the (u, v) path.
Notice that x occurs twice or zero times in our specified query range.
Therefore, the nodes which occur exactly once in this range are precisely those that are on the (u, v) path! (Try to convince yourself of why this is true : It's all because of dfs() properties.)

This forms the crux of our algorithm. While implementing Mo's, our add/remove function needs to check the number of times a particular node appears in a range. If it occurs twice (or zero times), then we don't take it's value into account! This can be easily implemented while moving the left and right pointers.

Case 2: P ≠ u

In this case, our query range would be [EN(u), ST(v)] + [ST(P), ST(P)].

The same logic as Case 1 applies here as well. The only difference is that we need to consider the value of P i.e the LCA separately, as it would not be counted in the query range.

This same problem is available on SPOJ.

If you aren't sure about some elements of this algorithm, take a look at this neat code.

Conclusion

We have effectively managed to reduce problem (2) to number of distinct values in a subarray by doing some careful bookkeeping. Now we can solve the problem in This modified DFS order works brilliantly to handle any type path queries and works well with Mo's algo. We can use a similar approach to solve many types of path query problems.

For example, consider the question of finding number of inversions in a (u, v) path in a Tree T, where each node has a value associated with it. This can now be solved in by using the above technique and maintaining a BIT or Segment Tree.

This is my first blog and I apologize for any mistakes that I may have made. I would like to thank sidhant for helping me understand this technique.

Sample Problems

Thanks a lot for reading! Comments (129)
 » 4 years ago, # | ← Rev. 2 →   In this case, our query range would be [EN(u), ST(v)] + [ST(P), ST(P)]. Shouldn't it be [EN(u), ST(v)] + [ST(P), ST(U)] or am I missing something ?
•  » » 4 years ago, # ^ | ← Rev. 2 →   Nope, it will be [ST(P), ST(P)].Consider the path from 3 to 5. In this case, P = 2EN(3) = 6 and ST(5) = 7, so we consider the range [6, 7] in A[] corresponding to the nodes [3, 5] giving us the values of nodes 3 and 5. Our query range does not consider the value of the lca as ST(P) < EN(u) < ST(v) < EN(P). Hence we must account for the value of P separately.
•  » » » Ok !!! Thanks Nice article anyways
•  » » » » Thank You :D
•  » » » » » I recently read an article which says that dynamic connectivity graph problem can be solved using using euler tour tree. can you write a blog on that please https://courses.csail.mit.edu/6.851/spring12/scribe/L20.pdf I cant figure a way to implement this
 » Thanks! That was a really nice tutorial!
•  » » Thanks a lot :)
 » Nice tutorial :)can you give links to some more problems on which similar approach can be applied ?
•  » »
•  » » » This can be done with standard Mo's Algorithm, because the queries are on subtrees and not paths.
•  » » » » Oh sorry, I thought this guy was asking for any problems related to the algorithms described above.
•  » » » » 2 years ago, # ^ | ← Rev. 2 →   but in each query, there is a new k. I wrote code for this problem, after whole implementation, i noticed that i missed the point that there is alway a new k for each query. Now i am not getting how can i solve this prob!my code!
•  » » » » » 2 years ago, # ^ | ← Rev. 2 →   You can create two maps,One that is updated on adding/removing the elements and it is keeping the counts on the Go. The other map is the map for MAP of K's like if element at position 2 was removed then MAP[count of 2]-- will be done that means there were x number of 2 but now there are x-1 of them so the occurrence of number that occurred x times (the contribution of 2 in x times is now reduced ) will be reduced by 1. Then in each query you will have to store the corresponding k's value as answer for that query.
 » 4 years ago, # | ← Rev. 2 →   BTW, we can find number of distinct values in a subarray [l, r] of a offline in O((q+n)\log n).Let's sort all queries by li.di = 1 if i is the first occurence of ai in a[l...n] otherwise di = 0.So, query (li, ri) is finding sum of d[l_i...r_i].When we move from query with l_i to query with l_i + 1 we must update only one or zero elements of di. It can be done in if we precalculated for each i next occurence of ai in array.
•  » » If you maintain the tree persistently, you can have an online solution as well.
•  » » » can you elaborate it . how to handle it online ? thanks in advance.
•  » » » » I think this is related to a current running contest.
•  » » » » » yes
•  » » » » » which contest ?
•  » » » » » » codechef feburary long challenge
•  » » » » » In curiosity i asked this question too early . sorry for that . you can answer it after contest is over
•  » » » » » » Yeah , 3 days early :P
•  » » » » » » » Contest has ended 3 months ago..Can anyone please answer this question now ?
•  » » » Can you please explain it how to do it online ?
•  » » » » Create an array of next occurences and build a persistent segment tree on that . The key idea is that number of distinct values in [L,R] is number of values whose next occurence is > R .
 » Thanks, Is known who used this idea on trees first time ?
•  » » It must have been known from before. But I guess this is the first proper tutorial/blog for it.
 » Totally went over my head! Excellent blog!
•  » » If it goes over your head, How do you realize it's an excellent ?
 » Thanks a lot you made my day !!I've been obsessing about COT2 for almost two months without anything that comes to mindif only i could upvote more than once
•  » » Thanks! I'm glad that you found it useful :)
 » I implemented this algorithm on the COT2 problem on SPOJ(http://www.spoj.com/problems/COT2/). I am getting WA. Can someone help me identify the bug in my code? http://ideone.com/aLS5Yx
•  » » I found the mistake. Thanks for the nice tutorial.
•  » » » What was the bug ?
•  » » » » Lol. Bro that was 7 months ago.
•  » » » » » Can you share your corrected code ? Because I'm getting a WA too .
•  » » » » » » 3 years ago, # ^ | ← Rev. 2 →   Sure. Link
•  » » » Can u please explain ur add and del functions. How are u maintaining the result after ignoring all those indexes which have occured 2 times?
•  » » » » Recently I solved one question using Mo's algorithm, and I remembered about this comment here. I overwrote the solution on the same link. Here is the solution for COT2. I think its self-explanatory how it is working.
 » 4 years ago, # | ← Rev. 2 →   Can someone please provide the algorithm for Problem 1?The best I could get is (N^2)*logN [as the sum of sizes of sets of each node is O(N^2) — Worst case linear graph with all values distinct]
•  » » 4 years ago, # ^ | ← Rev. 4 →   Maintain a set of values for each node in the tree. Let set(u) be the set of all values in the subtree rooted at u. We want size(set(u)) for all u.Let a node u have k children, v1, v2...vk. Every time you want to merge set(u) with set(vi), pop out the elements from the smaller set and insert them into the larger one. You can think of it like implementing union find, based on size. Consider any arbitrary node value. Every time you remove it from a certain set and insert it into some other, the size of the merged set is atleast twice the size of the original.Say you merge sets x and y. Assume size(x) ≤ size(y). Therefore, by the algorithm, you will push all the elements of x into y. Let xy be the merged set. size(xy) = size(x) + size(y). But size(y) ≥ size(x). So size(xy) ≥ 2 * size(x).Thus, each value will not move more than log n times. Since each move is done in O(logn), the total complexity for n values amounts to O(nlog2 n)Code
•  » » » Awesome! Thanks for the great explanation and code!
•  » » » 3 years ago, # ^ | ← Rev. 2 →   What if set x and set y isn't completely disjoint? In that case size(xy) = size(x) + size(y) statement isn't valid. Since the value on two nodes might be same.
•  » » » 3 years ago, # ^ | ← Rev. 3 →   I don't get the proof. Can you explain it a little more? size(xy) ≥ 2 * size(x) I think, size(xy) >= size(y), and size(xy) <= size(x) + size(y) Thus, each value will not move more than log n times. How? EDIT I think I understood. For a particular value to be included the maximum number of times in a move operation from set(x) to set(xy) where size(x) <= size(y), this value must be moved for each of it's ancestor upto root. That is only possible if the height of the tree is at most log n.But the size(xy) >= 2 * size(x) seems incorrect. I think you meant that the size of subtree of parent of x >= 2 * size(x).
•  » » » We cannot use the size function of the set to compare the sizes of the set as it would otherwise lead to N^2 complexity.Am i right?
•  » » » » set.size() is O(1).
•  » » » » » I thought it is O(n).Thanks for the info.
 » Very neatly written tutorial. You make it seem amazingly easy!
•  » » Thanks a lot :)
 » Superb idea! :DThanks! :D
 » Nicely written!
•  » » Thanks a lot :D
 » shouldn't it be end time of u to start time of v in case 1.If we start from start time of u then u will be included 2 times one for its start time and once for end time.Correct me if i am wrong..
•  » » Case 1 implies that u is an ancestor of v. Therefore, we won't visit u twice in the range [ST(u), ST(v)] as EN(u) > ST(v).
 » Has anyone managed to get accepted on the SPOJ problem with a Java solution? I'm getting NZEC Runtime Error, but it looks like it's actually due to time limit exceeding.
 » Update 1: Added sample problems.
•  » » For the "Frank Sinatra" problem. How could you find the less value not present in the path?I realize that any value greather than the size of the tree wouldn't change the answer. So, if i have at most 1E5 different values I can build a BIT. pos[i] = 1 if value i is present in the path. Then I binary search the less value k wich sum[0...k] is less than k. That would be my answer. However the complexity is O(N*sqrt(N)*log(N)*log(N)) and I think is excesive.
•  » » » The complexity wouldn't be , it would be . The first term is because you update your bit atmost times and the second term is because you binary search once for each query.
•  » » » » Thanks, my mistake. So, it is the best completely? Or there is another approach
•  » » » » » You can solve the problem in by doing square root decomposition on the values. Each update would be done in constant time and you will take additional time per query to find the block which has the smallest value. Code#include "bits/stdc++.h" using namespace std; typedef pair < int, int > pi; typedef vector < int > vi; typedef vector < pi > vpi; const int MAXN = 1e5 + 10; const int SQRT = 320; const int LOGN = 20; int A[2*MAXN]; pi B[MAXN]; int V[MAXN], C[MAXN]; int D[SQRT + 10][SQRT + 10], S[SQRT + 10]; int P[MAXN]; vpi G[MAXN]; int W[MAXN], R[MAXN]; struct query { int x; int y; int ind; }; query Q[MAXN]; bool comp(query a, query b) { if (int(a.x/SQRT) != int(b.x/SQRT)) return (int(a.x/SQRT) < int(b.x/SQRT)); return a.y < b.y; } int cont = 0; void dfs(int x, int p) { V[x] = 1; P[x] = p; A[cont] = x; B[x].first = cont; cont++; for (int i = 0; i < G[x].size(); ++i) { int y = G[x][i].first; int k = G[x][i].second; if (!V[y]) { C[y] = k; dfs(y, p + 1); } } A[cont] = x; B[x].second = cont; cont++; } void add(int x) { int k = C[A[x]]; if (k > MAXN) return; if (!W[A[x]]) { if (D[k/SQRT][k%SQRT] == 0) ++S[k/SQRT]; ++D[k/SQRT][k%SQRT]; ++W[A[x]]; } else { if (D[k/SQRT][k%SQRT] == 1) --S[k/SQRT]; --D[k/SQRT][k%SQRT]; --W[A[x]]; } } int main() { ios_base::sync_with_stdio(false); cin.tie(0); int n, m; cin >> n >> m; for (int i = 1; i < n; ++i) { int x, y, val; cin >> x >> y >> val; --x; --y; G[x].push_back(pi(y, val)); G[y].push_back(pi(x, val)); } dfs(0, 0); C = MAXN + 1; for (int i = 0; i < m; ++i) { int x, y; cin >> x >> y; --x; --y; if (B[x].first > B[y].first) swap(x, y); if (B[x].second > B[y].second) { Q[i].x = B[x].first + 1; Q[i].y = B[y].first; } else { Q[i].x = B[x].second; Q[i].y = B[y].first; } Q[i].ind = i; } sort(Q, Q + m, comp); int a = 0; int b = -1; for (int i = 0; i < m; ++i) { while (b < Q[i].y) { ++b; add(b); } while (a > Q[i].x) { --a; add(a); } while (b > Q[i].y) { add(b); --b; } while (a < Q[i].x) { add(a); ++a; } for (int w = 0; w < SQRT; ++w) { if (S[w] < SQRT) { for (int j = 0; j < SQRT; ++j) { if (D[w][j] == 0) { R[Q[i].ind] = w*SQRT + j; break; } } break; } } } for (int i = 0; i < m; ++i) cout << R[i] << endl; }
 » can it gives me tle,if i can't use weight compress?
•  » » If you do not compress weights, you'll need a map and that would add an additional log(n) factor. However, you might be able to squeeze your solution within the TL with an unordered_map.
 » BTW, there is a standard solution for the first problem (see this link in Russian). For each of the colors order all the vertices of this color according to the dfs traversal, let the vertices be labelled v1, v2, ..., vk. Add +1 to each of these verticies, and add -1 to the LCAs of the neighboring vertices lca(v1, v2), lca(v2, v3), ..., lca(vk - 1, vk). If you sum up the values inside a subtree, you get the number of distinct elements in it.Since the ordering can be done in O(n), and in theory you can answer lca queries for a static tree in O(1) with O(n) pre-processing, you have a linear solution (assuming 0 ≤ A[x] < N).
•  » » Thanks! This idea is pretty cool :)
•  » » » I understand the merging sets optimization!Can you explain how we can utilize this in Problem 1 (unique elements in subtree) to achieve ?As far as I see, lets say then we want, to change which can be done optimally when ,how do you propose to do it for ?
 » in problem frank sinatra, can i quickly find minimal number,that is not present in given set.
 » Isn't the time complexity of Mo's algorithm O(N*sqrt(Q)) instead of O(Q*sqrt(N))?
•  » » The complexity of Mo's depends on the number of times we increment/decrement the curL, curR variables. This link explains the time complexity of Mo's algorithm.
•  » » » 3 years ago, # ^ | ← Rev. 2 →   If the size of each block is k, then the time complexity of moving the left pointer is O(Q*k) and the time complexity of moving the right pointer is O(N/k*N). The optimal value of k is N/sqrt(Q) which results in total time complexity O(N*sqrt(Q)).
•  » » » » That is true. However, in most cases upper bounds on Q and N are equal (or pretty close), so it doesn't make a difference.
 » I wrote the code for COT2 judge gives runtime error at 10th testcase pls help me i cant find the error thanks in advancepls see my codehttps://ideone.com/MG3XbK
•  » » I think it's for the weight value. They haven't said anything about the limit. I compressed the weights.
 » Awesome Tutorial!
•  » » Thank you!
 » Can anyone explain how to linearize the tree .. (Not binary tree but any tree in general)As in Problem 1..
•  » »
•  » » » Thanks alot... !
 » I got Runtime Error. Here is my code There is any wrong my crealting lca tabel or anything else. Thanks in advance.
 » amazing tutorial!
 » 3 years ago, # | ← Rev. 3 →   If the tree store the values on the edges, you could store these values on the children (going from the root), and change the Case 2 to: if(P == u || P == v) check(P); .... asnwer the query if(P == u || P == v) check(P);
 » Why don't you write more tutorials?
 » http://www.spoj.com/problems/DQUERY/ a practice problem
 » COT2 code link doesnt work .
•  » » Updated.
 » Why does this get WA for COT2 :/ ?
 » Could you explain your idea for the problem of finding number of inversions in a (u, v) path in a Tree T.
•  » » Just maintain a BIT during Mo's
 » 3 years ago, # | ← Rev. 2 → If I flatten the above tree, my array would be:8 3 1 6 4 7 10 14 13Suppose I need to use Mo's algorithm for subtrees(assume I need to find sum of values of each subtree indicated by the query)For a given query 'Vj' how would I find its end range index in the array?Eg if given query is node '6', the starting range would be idx 3 and ending would be idx 5.
•  » » Store the starting and ending times for every node during your dfs .
•  » » » Thanks, understood it after a bit of googling about discovery/begin/end times.
 » For a problem like this: http://lightoj.com/volume_showproblem.php?problem=1348 where I need to return sum of all the nodes in a given path & update the value of a node, how should I approach using this technique of linearizing the tree? I mean since I need to ignore nodes which have occurrence of 2 so the range becomes discontinuous for a segment tree structure.
•  » » I think you should use Heavy light decomposition(HLD).https://wcipeg.com/wiki/Heavy-light_decomposition
 » Nice Article ... BTW Can we solve the problem with Binary lifting ?
 » How can apply this method if weight is given on edges instead vertices
•  » » Root the tree arbitarily . Map the weight of edge (parent-child) to the child .
•  » » » and what about the weight of root?
•  » » » » Are you saying that both edges and nodes have weights ?
•  » » » » » no, i mean as you said map the weight of edge (parent-child) to the child but root has no parent so what value is map to the root of the tree?
•  » » » » » » nothing is mapped to the root . A tree has n — 1 edges which would be mapped to n — 1 vertices of the tree .
•  » » » » » » » thanks now i understand
•  » » » » » » assign it an impossible value like -INFINITY. This means whenever you see this value, you know it's not allowed, and you ignore it.
 » how to calculate LCA fast? i only know the O(n) algorithm...
•  » »
 » In this case, our query range would be[EN(u), ST(v)] + [ST(P), ST(P)].Consider on this case, if we select 3 and 8 on the tree given to explain the DFS-Order, the range[EN(u), ST(v)] contains the whole subtree S(5) which is not on our query path. Are we supposed to judge every node in the range or Am I missing something? Thx!
•  » » you only consider nodes which appear once in the range, so maintain a frequency count of the current nodes, if a node appears twice, remove it from the list of nodes
 » Thanks a lot, but I don't really understand the meaning of If it occurs twice (or zero times), then we don't take it's value into account! This can be easily implemented while moving the left and right pointers. and I wrote the code as what you have said in this essay,but it seems wrong and I don't know where I count the answer incorrectly my code
 » Man, you nailed the use DFS. BTW thanx for the amazing article.
 » "One easy way to solve this is to flatten the tree into an array by doing a Preorder traversal" Isn't Preorder traversal done on a binary tree? The given tree may not be a binary tree. Where am I going wrong?
•  » » By Preorder, I meant that we look at the value of v before looking at the value of all of its descendants.
 » How would we handle nodes which occurs two times because they were not in the required path ?In other words how would we remove their contribution ?
•  » » 2 years ago, # ^ | ← Rev. 2 →   keep separate counter for counting occurences of element in present range !!while changing L and R,if occurence of element becomes 1 from 2 then consider that element as ADDED if occurences of element becomes 0 from 1 then consider that element REMOVED if occurences of element becomes 2 from 1 then consider that element REMOVED if occurences of element becomes 1 form 0 then consider that element ADDED
 » How can we use this modified dfs technique for weighted edge Problems like http://www.spoj.com/problems/QTREE/.
 » Is the idea of converting a tree into a linear array is applicable on segment tree? If applicable then how to handle which node appear 2 times
•  » » You cannot really do it. You should use heavy light decomposition.
 » Please if anyone can explain the struct part in the code mentioned in the blog. Here's the code snippet struct query { int id,l,r,lc; bool operator < (const query& rhs) { return (BL[l] == BL[rhs.l]) ? (r
•  » » To sort queries we need to know how to compare them, it can be done by overloading operator < or making a bool function (comparator) which takes two queries and return true if first must be to the left of the second in sorted array.
•  » » » 2 years ago, # ^ | ← Rev. 2 →   Thank you for explaining :-)It is sorting the l values according to the BLOCK value of l i.e. BL[l]. If both the values are same then it compares the one which has smaller r value just like in Standard Mo's algorithm. Am I interpreting it right?
•  » » » » Yeah, firstly we compare blocks of left edge and compare right edge if they are equal.
 » 2 years ago, # | ← Rev. 3 →   very nicely explained. thanks a lot rekt_n00b this article is very helpful
 » 2 years ago, # | ← Rev. 4 →   In this case, our query range would be [EN(u), ST(v)] + [ST(P), ST(P)].Is the range went wrong. For example: If we modify dfs-order then we will have : 1-2-3-4-5-6-7-8-9-10-11. We have a query count distinct number in unique path from 4 to 11. Then the answer contain S(5) — out of path from 4 to 11 ?? Am I missing something ?
•  » » i think DFS-order will be 1-2-3-3-4-4-2-5-6-6-7-7-5-8-9-9-10-10-11-11-8-1, EN=6, ST=19, so the nodes which occur exactly once in this range are 4,2,8,11; then you have to consider also ST=1 .In the end you have 4,2,8,11,1
•  » » » Thank you.
 » This article really helped me a lot! I wrote a similar tutorial for euler tour + segment tree with reference to the problem OTOCI on SPOJ. Check it out if you're interested! https://t00pr0.wordpress.com/2018/01/04/spoj-otoci-solution-and-a-tutorial-on-flattening-trees-using-euler-order/
 » Can anyone tell me how to solve Vasya and little Bear problem (3rd problem) ??
 » What is the use of VAL[] in the ideone link provided ? I mean i understand the use of A[x] that stores the position of x in sorted weights..
 » Nice tutorial!! A problem using this concept 852I - Dating.
 » rekt_n00b thanks a lot i badly needed this!!!!
 » 4 months ago, # | ← Rev. 2 →   In case any body isn't clear about how to transform the tree- I run a simple DFS, save them in an N+N sized array, by the order of starting & finishing time. Then ordered the queries offline by the given node's starting and ending time. Just run the query by MO's. I love this problem.