animeshf's blog

By animeshf, history, 8 years ago,

Introduction

Mo's Algorithm has become pretty popular in the past few years and is now considered as a pretty standard technique in the world of Competitive Programming. This blog will describe a method to generalize Mo's algorithm to maintain information about paths between nodes in a tree.

Prerequisites

Mo's Algorithm — If you do not know this yet, read this amazing article before continuing with this blog.

Preorder Traversal or DFS Order of the Tree.

Problem 1 — Handling Subtree Queries

Consider the following problem. You will be given a rooted Tree T of N nodes where each node is associated with a value A[node]. You need to handle Q queries, each comprising one integer u. In each query you must report the number of distinct values in the subtree rooted at u. In other words, if you store all the values in the subtree rooted at u in a set, what would be the size of this set?

Constraints

1 ≤ N, Q ≤ 105

1 ≤ A[node] ≤ 109

Solution(s)

Seems pretty simple, doesn't it? One easy way to solve this is to flatten the tree into an array by doing a Preorder traversal and then implement Mo's Algorithm. Maintain a lookup table which maintains the frequency of each value in the current window. By maintaining this, the answer can be updated easily. The complexity of this algorithm would be

Note that you can also solve this in by maintaining a set in each node and merging the smaller set into the larger ones.

Problem 2 — Handling Path Queries

Now let's modify Problem 1 a little. Instead of computing the number of distinct values in a subtree, compute the number of distinct values in the unique path from u to v. I recommend you to pause here and try solving the problem for a while. The constraints of this problem are the same as Problem 1.

The Issue

An important reason why Problem (1) worked beautifully was because the dfs-order traversal made it possible to represent any subtree as a contiguous range in an array. Thus the problem was reduced to "finding number of distinct values in a subarray [L, R] of A[]. Note that it is not possible to do so for path queries, as nodes which are O(N) distance apart in the tree might be O(1) distance apart in the flattened tree (represented by Array A[]). So doing a normal dfs-order would not work out.

Observation(s)

Let a node u have k children. Let us number them as v1,v2...vk. Let S(u) denote the subtree rooted at u.

Let us assume that dfs() will visit u's children in the order v1,v2...vk. Let x be any node in S(vi) and y be any node in S(vj) and let i < j. Notice that dfs(y) will be called only after dfs(x) has been completed and S(x) has been explored. Thus, before we call dfs(y), we would have entered and exited S(x). We will exploit this seemingly obvious property of dfs() to modify our existing algorithm and try to represent each query as a contiguous range in a flattened array.

Modified DFS-Order

Let us modify the dfs order as follows. For each node u, maintain the Start and End time of S(u). Let's call them ST(u) and EN(u). The only change you need to make is that you must increment the global timekeeping variable even when you finish traversing some subtree (EN(u) = ++cur). In short, we will maintain 2 values for each node u. One will denote the time when you entered S(u) and the other would denote the time when you exited S(u). Consider the tree in the picture. Given below are the ST() and EN() values of the nodes.

ST(1) = 1 EN(1) = 18

ST(2) = 2 EN(2) = 11

ST(3) = 3 EN(3) = 6

ST(4) = 4 EN(4) = 5

ST(5) = 7 EN(5) = 10

ST(6) = 8 EN(6) = 9

ST(7) = 12 EN(7) = 17

ST(8) = 13 EN(8) = 14

ST(9) = 15 EN(9) = 16

A[] = {1, 2, 3, 4, 4, 3, 5, 6, 6, 5, 2, 7, 8, 8, 9, 9, 7, 1}

The Algorithm

Now that we're equipped with the necessary weapons, let's understand how to process the queries.

Let a query be (u, v). We will try to map each query to a range in the flattened array. Let ST(u) ≤ ST(v) where ST(u) denotes visit time of node u in T. Let P = LCA(u, v) denote the lowest common ancestor of nodes u and v. There are 2 possible cases:

Case 1: P = u

In this case, our query range would be [ST(u), ST(v)]. Why will this work?

Consider any node x that does not lie in the (u, v) path.
Notice that x occurs twice or zero times in our specified query range.
Therefore, the nodes which occur exactly once in this range are precisely those that are on the (u, v) path! (Try to convince yourself of why this is true : It's all because of dfs() properties.)

This forms the crux of our algorithm. While implementing Mo's, our add/remove function needs to check the number of times a particular node appears in a range. If it occurs twice (or zero times), then we don't take it's value into account! This can be easily implemented while moving the left and right pointers.

Case 2: P ≠ u

In this case, our query range would be [EN(u), ST(v)] + [ST(P), ST(P)].

The same logic as Case 1 applies here as well. The only difference is that we need to consider the value of P i.e the LCA separately, as it would not be counted in the query range.

This same problem is available on SPOJ.

If you aren't sure about some elements of this algorithm, take a look at this neat code.

Conclusion

We have effectively managed to reduce problem (2) to number of distinct values in a subarray by doing some careful bookkeeping. Now we can solve the problem in This modified DFS order works brilliantly to handle any type path queries and works well with Mo's algo. We can use a similar approach to solve many types of path query problems.

For example, consider the question of finding number of inversions in a (u, v) path in a Tree T, where each node has a value associated with it. This can now be solved in by using the above technique and maintaining a BIT or Segment Tree.

This is my first blog and I apologize for any mistakes that I may have made. I would like to thank sidhant for helping me understand this technique.

Sample Problems

• +407

| Write comment?
 » 8 years ago, # | ← Rev. 2 →   +8 In this case, our query range would be [EN(u), ST(v)] + [ST(P), ST(P)]. Shouldn't it be [EN(u), ST(v)] + [ST(P), ST(U)] or am I missing something ?
•  » » 8 years ago, # ^ | ← Rev. 2 →   +8 Nope, it will be [ST(P), ST(P)].Consider the path from 3 to 5. In this case, P = 2EN(3) = 6 and ST(5) = 7, so we consider the range [6, 7] in A[] corresponding to the nodes [3, 5] giving us the values of nodes 3 and 5. Our query range does not consider the value of the lca as ST(P) < EN(u) < ST(v) < EN(P). Hence we must account for the value of P separately.
•  » » » 8 years ago, # ^ |   +8 Ok !!! Thanks Nice article anyways
•  » » » » 8 years ago, # ^ |   +13 Thank You :D
 » 8 years ago, # |   +23 Thanks! That was a really nice tutorial!
•  » » 8 years ago, # ^ |   +13 Thanks a lot :)
 » 8 years ago, # |   0 Nice tutorial :)can you give links to some more problems on which similar approach can be applied ?
•  » » 8 years ago, # ^ |   0
•  » » » 8 years ago, # ^ |   0 This can be done with standard Mo's Algorithm, because the queries are on subtrees and not paths.
•  » » » » 8 years ago, # ^ |   0 Oh sorry, I thought this guy was asking for any problems related to the algorithms described above.
•  » » » » 7 years ago, # ^ | ← Rev. 2 →   0 but in each query, there is a new k. I wrote code for this problem, after whole implementation, i noticed that i missed the point that there is alway a new k for each query. Now i am not getting how can i solve this prob!my code!
 » 8 years ago, # | ← Rev. 2 →   +5 BTW, we can find number of distinct values in a subarray [l, r] of a offline in O((q+n)\log n).Let's sort all queries by li.di = 1 if i is the first occurence of ai in a[l...n] otherwise di = 0.So, query (li, ri) is finding sum of d[l_i...r_i].When we move from query with l_i to query with l_i + 1 we must update only one or zero elements of di. It can be done in if we precalculated for each i next occurence of ai in array.
•  » » 8 years ago, # ^ |   +10 If you maintain the tree persistently, you can have an online solution as well.
•  » » » 7 years ago, # ^ |   -6 can you elaborate it . how to handle it online ? thanks in advance.
•  » » » » 7 years ago, # ^ |   0 I think this is related to a current running contest.
•  » » » » » 7 years ago, # ^ |   -6 yes
•  » » » » » 7 years ago, # ^ |   0 which contest ?
•  » » » » » » 7 years ago, # ^ |   0 codechef feburary long challenge
•  » » » » » 7 years ago, # ^ |   0 In curiosity i asked this question too early . sorry for that . you can answer it after contest is over
•  » » » » » » 7 years ago, # ^ |   +1 Yeah , 3 days early :P
•  » » » » » » » 7 years ago, # ^ |   0 Contest has ended 3 months ago..Can anyone please answer this question now ?
•  » » » 7 years ago, # ^ |   0 Can you please explain it how to do it online ?
•  » » » » 7 years ago, # ^ |   -7 Create an array of next occurences and build a persistent segment tree on that . The key idea is that number of distinct values in [L,R] is number of values whose next occurence is > R .
 » 8 years ago, # |   0 Thanks, Is known who used this idea on trees first time ?
•  » » 8 years ago, # ^ |   +5 It must have been known from before. But I guess this is the first proper tutorial/blog for it.
 » 8 years ago, # |   +1 Totally went over my head! Excellent blog!
•  » » 8 years ago, # ^ |   +3 If it goes over your head, How do you realize it's an excellent ?
 » 8 years ago, # |   0 Thanks a lot you made my day !!I've been obsessing about COT2 for almost two months without anything that comes to mindif only i could upvote more than once
•  » » 8 years ago, # ^ |   +1 Thanks! I'm glad that you found it useful :)
 » 8 years ago, # |   0 I implemented this algorithm on the COT2 problem on SPOJ(http://www.spoj.com/problems/COT2/). I am getting WA. Can someone help me identify the bug in my code? http://ideone.com/aLS5Yx
•  » » 8 years ago, # ^ |   +8 I found the mistake. Thanks for the nice tutorial.
•  » » » 8 years ago, # ^ |   0 What was the bug ?
•  » » » » 8 years ago, # ^ |   0 Lol. Bro that was 7 months ago.
•  » » » » » 8 years ago, # ^ |   0 Can you share your corrected code ? Because I'm getting a WA too .
•  » » » » » » 8 years ago, # ^ | ← Rev. 2 →   0 Sure. Link
•  » » » 8 years ago, # ^ |   0 Can u please explain ur add and del functions. How are u maintaining the result after ignoring all those indexes which have occured 2 times?
•  » » » » 8 years ago, # ^ |   0 Recently I solved one question using Mo's algorithm, and I remembered about this comment here. I overwrote the solution on the same link. Here is the solution for COT2. I think its self-explanatory how it is working.
 » 8 years ago, # | ← Rev. 2 →   0 Can someone please provide the algorithm for Problem 1?The best I could get is (N^2)*logN [as the sum of sizes of sets of each node is O(N^2) — Worst case linear graph with all values distinct]
•  » » 8 years ago, # ^ | ← Rev. 4 →   +3 Maintain a set of values for each node in the tree. Let set(u) be the set of all values in the subtree rooted at u. We want size(set(u)) for all u.Let a node u have k children, v1, v2...vk. Every time you want to merge set(u) with set(vi), pop out the elements from the smaller set and insert them into the larger one. You can think of it like implementing union find, based on size. Consider any arbitrary node value. Every time you remove it from a certain set and insert it into some other, the size of the merged set is atleast twice the size of the original.Say you merge sets x and y. Assume size(x) ≤ size(y). Therefore, by the algorithm, you will push all the elements of x into y. Let xy be the merged set. size(xy) = size(x) + size(y). But size(y) ≥ size(x). So size(xy) ≥ 2 * size(x).Thus, each value will not move more than log n times. Since each move is done in O(logn), the total complexity for n values amounts to O(nlog2 n)Code
•  » » » 8 years ago, # ^ |   +5 Awesome! Thanks for the great explanation and code!
•  » » » 8 years ago, # ^ | ← Rev. 3 →   0 I don't get the proof. Can you explain it a little more? size(xy) ≥ 2 * size(x) I think, size(xy) >= size(y), and size(xy) <= size(x) + size(y) Thus, each value will not move more than log n times. How? EDIT I think I understood. For a particular value to be included the maximum number of times in a move operation from set(x) to set(xy) where size(x) <= size(y), this value must be moved for each of it's ancestor upto root. That is only possible if the height of the tree is at most log n.But the size(xy) >= 2 * size(x) seems incorrect. I think you meant that the size of subtree of parent of x >= 2 * size(x).
•  » » » 8 years ago, # ^ |   0 We cannot use the size function of the set to compare the sizes of the set as it would otherwise lead to N^2 complexity.Am i right?
•  » » » » 8 years ago, # ^ |   +6 set.size() is O(1).
•  » » » » » 8 years ago, # ^ |   0 I thought it is O(n).Thanks for the info.
 » 8 years ago, # |   0 Very neatly written tutorial. You make it seem amazingly easy!
•  » » 8 years ago, # ^ |   0 Thanks a lot :)
 » 8 years ago, # |   +11 Superb idea! :DThanks! :D
 » 8 years ago, # |   0 Nicely written!
•  » » 8 years ago, # ^ |   0 Thanks a lot :D
 » 8 years ago, # |   0 shouldn't it be end time of u to start time of v in case 1.If we start from start time of u then u will be included 2 times one for its start time and once for end time.Correct me if i am wrong..
•  » » 8 years ago, # ^ |   0 Case 1 implies that u is an ancestor of v. Therefore, we won't visit u twice in the range [ST(u), ST(v)] as EN(u) > ST(v).
 » 8 years ago, # |   0 Has anyone managed to get accepted on the SPOJ problem with a Java solution? I'm getting NZEC Runtime Error, but it looks like it's actually due to time limit exceeding.
 » 8 years ago, # |   +3 Update 1: Added sample problems.
•  » » 8 years ago, # ^ |   0 For the "Frank Sinatra" problem. How could you find the less value not present in the path?I realize that any value greather than the size of the tree wouldn't change the answer. So, if i have at most 1E5 different values I can build a BIT. pos[i] = 1 if value i is present in the path. Then I binary search the less value k wich sum[0...k] is less than k. That would be my answer. However the complexity is O(N*sqrt(N)*log(N)*log(N)) and I think is excesive.
•  » » » 8 years ago, # ^ |   0 The complexity wouldn't be , it would be . The first term is because you update your bit atmost times and the second term is because you binary search once for each query.
•  » » » » 8 years ago, # ^ |   0 Thanks, my mistake. So, it is the best completely? Or there is another approach
•  » » » » » 8 years ago, # ^ |   0 You can solve the problem in by doing square root decomposition on the values. Each update would be done in constant time and you will take additional time per query to find the block which has the smallest value. Code#include "bits/stdc++.h" using namespace std; typedef pair < int, int > pi; typedef vector < int > vi; typedef vector < pi > vpi; const int MAXN = 1e5 + 10; const int SQRT = 320; const int LOGN = 20; int A[2*MAXN]; pi B[MAXN]; int V[MAXN], C[MAXN]; int D[SQRT + 10][SQRT + 10], S[SQRT + 10]; int P[MAXN]; vpi G[MAXN]; int W[MAXN], R[MAXN]; struct query { int x; int y; int ind; }; query Q[MAXN]; bool comp(query a, query b) { if (int(a.x/SQRT) != int(b.x/SQRT)) return (int(a.x/SQRT) < int(b.x/SQRT)); return a.y < b.y; } int cont = 0; void dfs(int x, int p) { V[x] = 1; P[x] = p; A[cont] = x; B[x].first = cont; cont++; for (int i = 0; i < G[x].size(); ++i) { int y = G[x][i].first; int k = G[x][i].second; if (!V[y]) { C[y] = k; dfs(y, p + 1); } } A[cont] = x; B[x].second = cont; cont++; } void add(int x) { int k = C[A[x]]; if (k > MAXN) return; if (!W[A[x]]) { if (D[k/SQRT][k%SQRT] == 0) ++S[k/SQRT]; ++D[k/SQRT][k%SQRT]; ++W[A[x]]; } else { if (D[k/SQRT][k%SQRT] == 1) --S[k/SQRT]; --D[k/SQRT][k%SQRT]; --W[A[x]]; } } int main() { ios_base::sync_with_stdio(false); cin.tie(0); int n, m; cin >> n >> m; for (int i = 1; i < n; ++i) { int x, y, val; cin >> x >> y >> val; --x; --y; G[x].push_back(pi(y, val)); G[y].push_back(pi(x, val)); } dfs(0, 0); C[0] = MAXN + 1; for (int i = 0; i < m; ++i) { int x, y; cin >> x >> y; --x; --y; if (B[x].first > B[y].first) swap(x, y); if (B[x].second > B[y].second) { Q[i].x = B[x].first + 1; Q[i].y = B[y].first; } else { Q[i].x = B[x].second; Q[i].y = B[y].first; } Q[i].ind = i; } sort(Q, Q + m, comp); int a = 0; int b = -1; for (int i = 0; i < m; ++i) { while (b < Q[i].y) { ++b; add(b); } while (a > Q[i].x) { --a; add(a); } while (b > Q[i].y) { add(b); --b; } while (a < Q[i].x) { add(a); ++a; } for (int w = 0; w < SQRT; ++w) { if (S[w] < SQRT) { for (int j = 0; j < SQRT; ++j) { if (D[w][j] == 0) { R[Q[i].ind] = w*SQRT + j; break; } } break; } } } for (int i = 0; i < m; ++i) cout << R[i] << endl; } 
•  » » » » » » 5 years ago, # ^ | ← Rev. 3 →   0 In your code, I saw this: if (B[x].first > B[y].first) swap(x, y); if (B[x].second > B[y].second) { Q[i].x = B[x].first + 1; Q[i].y = B[y].first; } else { Q[i].x = B[x].second; Q[i].y = B[y].first; } Q[i].ind = i; Can " if (B[x].second > B[y].second) " replace LCA(x, y) == x ?I also want to ask does the O( N * sqrt(N) * log(N) + N * log (N) ) algorithm pass this problem ?
 » 8 years ago, # |   0 can it gives me tle,if i can't use weight compress?
•  » » 8 years ago, # ^ |   0 If you do not compress weights, you'll need a map and that would add an additional log(n) factor. However, you might be able to squeeze your solution within the TL with an unordered_map.
 » 8 years ago, # |   +39 BTW, there is a standard solution for the first problem (see this link in Russian). For each of the colors order all the vertices of this color according to the dfs traversal, let the vertices be labelled v1, v2, ..., vk. Add +1 to each of these verticies, and add -1 to the LCAs of the neighboring vertices lca(v1, v2), lca(v2, v3), ..., lca(vk - 1, vk). If you sum up the values inside a subtree, you get the number of distinct elements in it.Since the ordering can be done in O(n), and in theory you can answer lca queries for a static tree in O(1) with O(n) pre-processing, you have a linear solution (assuming 0 ≤ A[x] < N).
•  » » 8 years ago, # ^ |   0 Thanks! This idea is pretty cool :)
•  » » » 7 years ago, # ^ |   +3 I understand the merging sets optimization!Can you explain how we can utilize this in Problem 1 (unique elements in subtree) to achieve ?As far as I see, lets say then we want, to change which can be done optimally when ,how do you propose to do it for ?
 » 8 years ago, # |   +5 Isn't the time complexity of Mo's algorithm O(N*sqrt(Q)) instead of O(Q*sqrt(N))?
•  » » 8 years ago, # ^ |   0 The complexity of Mo's depends on the number of times we increment/decrement the curL, curR variables. This link explains the time complexity of Mo's algorithm.
•  » » » 8 years ago, # ^ | ← Rev. 2 →   +10 If the size of each block is k, then the time complexity of moving the left pointer is O(Q*k) and the time complexity of moving the right pointer is O(N/k*N). The optimal value of k is N/sqrt(Q) which results in total time complexity O(N*sqrt(Q)).
•  » » » » 8 years ago, # ^ |   0 That is true. However, in most cases upper bounds on Q and N are equal (or pretty close), so it doesn't make a difference.
 » 8 years ago, # |   0 I wrote the code for COT2 judge gives runtime error at 10th testcase pls help me i cant find the error thanks in advancepls see my codehttps://ideone.com/MG3XbK
•  » » 5 years ago, # ^ |   0 I think it's for the weight value. They haven't said anything about the limit. I compressed the weights.
 » 8 years ago, # |   +19 Awesome Tutorial!
•  » » 8 years ago, # ^ |   +3 Thank you!
 » 8 years ago, # |   0 Can anyone explain how to linearize the tree .. (Not binary tree but any tree in general)As in Problem 1..
•  » » 8 years ago, # ^ |   0
 » 8 years ago, # |   0 I got Runtime Error. Here is my code There is any wrong my crealting lca tabel or anything else. Thanks in advance.
 » 8 years ago, # |   0 amazing tutorial!
 » 8 years ago, # | ← Rev. 3 →   0 If the tree store the values on the edges, you could store these values on the children (going from the root), and change the Case 2 to: if(P == u || P == v) check(P); .... asnwer the query if(P == u || P == v) check(P); 
 » 8 years ago, # |   0 Why don't you write more tutorials?
 » 8 years ago, # |   +3 COT2 code link doesnt work .
•  » » 8 years ago, # ^ |   +7 Updated.
 » 8 years ago, # |   0 Why does this get WA for COT2 :/ ?
 » 8 years ago, # |   +5 Could you explain your idea for the problem of finding number of inversions in a (u, v) path in a Tree T.
•  » » 7 years ago, # ^ |   0 Just maintain a BIT during Mo's
 » 8 years ago, # | ← Rev. 2 →   0 If I flatten the above tree, my array would be:8 3 1 6 4 7 10 14 13Suppose I need to use Mo's algorithm for subtrees(assume I need to find sum of values of each subtree indicated by the query)For a given query 'Vj' how would I find its end range index in the array?Eg if given query is node '6', the starting range would be idx 3 and ending would be idx 5.
•  » » 8 years ago, # ^ |   +1 Store the starting and ending times for every node during your dfs .
•  » » » 8 years ago, # ^ |   0 Thanks, understood it after a bit of googling about discovery/begin/end times.
 » 8 years ago, # |   0 For a problem like this: http://lightoj.com/volume_showproblem.php?problem=1348 where I need to return sum of all the nodes in a given path & update the value of a node, how should I approach using this technique of linearizing the tree? I mean since I need to ignore nodes which have occurrence of 2 so the range becomes discontinuous for a segment tree structure.
•  » » 7 years ago, # ^ |   0 I think you should use Heavy light decomposition(HLD).https://wcipeg.com/wiki/Heavy-light_decomposition
•  » » » 5 years ago, # ^ | ← Rev. 3 →   0 It can be done without HLD. Let be dp[u] the sum of values of path from root downto u. The sum of values on the path from u to v is dp[u]+dp[v]-2*dp[L]+val[L] where L is lca(u,v). When you change the value of a node u, you must change dp[v] for all v belonging to subtree of u. With an euler tour the subtree of a node becomes a continous subarray, so you can easily update it with a segment tree or similars.
 » 8 years ago, # |   0 How can apply this method if weight is given on edges instead vertices
•  » » 8 years ago, # ^ |   0 Root the tree arbitarily . Map the weight of edge (parent-child) to the child .
•  » » » 8 years ago, # ^ |   0 and what about the weight of root?
•  » » » » 8 years ago, # ^ |   0 Are you saying that both edges and nodes have weights ?
•  » » » » » 8 years ago, # ^ |   0 no, i mean as you said map the weight of edge (parent-child) to the child but root has no parent so what value is map to the root of the tree?
•  » » » » » » 8 years ago, # ^ |   0 nothing is mapped to the root . A tree has n — 1 edges which would be mapped to n — 1 vertices of the tree .
•  » » » » » » 8 years ago, # ^ |   0 assign it an impossible value like -INFINITY. This means whenever you see this value, you know it's not allowed, and you ignore it.
 » 8 years ago, # |   0 how to calculate LCA fast? i only know the O(n) algorithm...
•  » » 8 years ago, # ^ |   +5
 » 8 years ago, # |   0 In this case, our query range would be[EN(u), ST(v)] + [ST(P), ST(P)].Consider on this case, if we select 3 and 8 on the tree given to explain the DFS-Order, the range[EN(u), ST(v)] contains the whole subtree S(5) which is not on our query path. Are we supposed to judge every node in the range or Am I missing something? Thx!
•  » » 8 years ago, # ^ |   0 you only consider nodes which appear once in the range, so maintain a frequency count of the current nodes, if a node appears twice, remove it from the list of nodes
 » 7 years ago, # |   +4 Man, you nailed the use DFS. BTW thanx for the amazing article.
 » 7 years ago, # |   0 "One easy way to solve this is to flatten the tree into an array by doing a Preorder traversal" Isn't Preorder traversal done on a binary tree? The given tree may not be a binary tree. Where am I going wrong?
•  » » 7 years ago, # ^ |   +6 By Preorder, I meant that we look at the value of v before looking at the value of all of its descendants.
 » 7 years ago, # |   0 How would we handle nodes which occurs two times because they were not in the required path ?In other words how would we remove their contribution ?
•  » » 7 years ago, # ^ | ← Rev. 2 →   0 keep separate counter for counting occurences of element in present range !!while changing L and R,if occurence of element becomes 1 from 2 then consider that element as ADDED if occurences of element becomes 0 from 1 then consider that element REMOVED if occurences of element becomes 2 from 1 then consider that element REMOVED if occurences of element becomes 1 form 0 then consider that element ADDED
 » 7 years ago, # |   0 Is the idea of converting a tree into a linear array is applicable on segment tree? If applicable then how to handle which node appear 2 times
•  » » 7 years ago, # ^ |   0 You cannot really do it. You should use heavy light decomposition.
 » 7 years ago, # |   0 Please if anyone can explain the struct part in the code mentioned in the blog. Here's the code snippet struct query { int id,l,r,lc; bool operator < (const query& rhs) { return (BL[l] == BL[rhs.l]) ? (r
•  » » 7 years ago, # ^ |   +1 To sort queries we need to know how to compare them, it can be done by overloading operator < or making a bool function (comparator) which takes two queries and return true if first must be to the left of the second in sorted array.
•  » » » 7 years ago, # ^ | ← Rev. 2 →   0 Thank you for explaining :-)It is sorting the l values according to the BLOCK value of l i.e. BL[l]. If both the values are same then it compares the one which has smaller r value just like in Standard Mo's algorithm. Am I interpreting it right?
•  » » » » 7 years ago, # ^ |   +1 Yeah, firstly we compare blocks of left edge and compare right edge if they are equal.
 » 7 years ago, # | ← Rev. 4 →   0 In this case, our query range would be [EN(u), ST(v)] + [ST(P), ST(P)].Is the range went wrong. For example:If we modify dfs-order then we will have : 1-2-3-4-5-6-7-8-9-10-11. We have a query count distinct number in unique path from 4 to 11. Then the answer contain S(5) — out of path from 4 to 11 ?? Am I missing something ?
•  » » 7 years ago, # ^ |   0 i think DFS-order will be 1-2-3-3-4-4-2-5-6-6-7-7-5-8-9-9-10-10-11-11-8-1, EN[4]=6, ST[11]=19, so the nodes which occur exactly once in this range are 4,2,8,11; then you have to consider also ST[1]=1 .In the end you have 4,2,8,11,1
 » 6 years ago, # |   0 Can anyone tell me how to solve Vasya and little Bear problem (3rd problem) ??
 » 6 years ago, # |   -8 What is the use of VAL[] in the ideone link provided ? I mean i understand the use of A[x] that stores the position of x in sorted weights..
 » 5 years ago, # |   0 Nice tutorial!! A problem using this concept 852I - Dating.
 » 5 years ago, # | ← Rev. 2 →   0 In case any body isn't clear about how to transform the tree- I run a simple DFS, save them in an N+N sized array, by the order of starting & finishing time. Then ordered the queries offline by the given node's starting and ending time. Just run the query by MO's. I love this problem.
 » 5 years ago, # |   0 Can someone explain that inversion count example in the tutorial as it's unclear to me? Also, can a problem link be given for that?
 » 5 years ago, # | ← Rev. 2 →   0 [Deleted]
 » 5 years ago, # |   0 Thanks!! This blog was extremely helpful for me.
 » 4 years ago, # |   0 please explain the use of const int LN = 17 in your code
•  » » 4 years ago, # ^ | ← Rev. 6 →   +4 You mean LN=19 rather? It seems to be used for setting the size of the first dimension of the sparse table. Because 2^19 is quite a huge value (> 500K). So the author probably doesn't need to worry about running out of bounds. You know that one dimension of the sparse table has to be log(N) where N is the size of the array you are computing over right? So LN must be stated such such that 2^LN > N.
 » 4 years ago, # |   +5 A problem in April LOng Appears om this concept : link is FCTRE
 » 4 years ago, # | ← Rev. 2 →   +1 Thanks for the amazing tutorial !
 » 4 years ago, # |   0 Great tutorial ! I wanted to know, if MOs can be applied, when edges are weighted and nodes don't have any value. In some of the cases, like sum of edges on path, I thought of distributing the weights of edges on the nodes , which are connected by the edge. So we can sum up the nodes + add an extra u,v and then dividing it by 2. Is there a general way to approach such type of queries ? Thanks
 » 4 years ago, # |   0 I tried to solve the Problem 1 using Segment Tree but it is giving TLE Problem LinkI think merging two sets is taking linear time How do I optimize it?
 » 4 years ago, # |   0 Nice algorithm and blog
 » 3 years ago, # |   0 How to find the LCA of two nodes here.. is there any simpler way than doing binary lifting??
 » 3 years ago, # |   +3 where can i submit code for Frank Sinatra — Problem F ?
•  » » 3 years ago, # ^ |   +16 You can submit it in the gym contest it was taken from.
•  » » 20 months ago, # ^ |   0 Did you get AC in it ? If yes, can you please share your code ?
 » 20 months ago, # |   0 How to solve Vasya and little bear ?
 » 16 months ago, # |   0 Thanks for sharing. It is amazing! - In case I: It says [ST[u], ST[v]], but couldn't it be [EN[u], ST[v]]? In Problem 2 — Handling Path Queries, A[] = {1, 2, 3, 4, 4, 3, 5, 6, 6, 5, 2, 7, 8, 8, 9, 9, 7, 1} Let's says the query is from node 4 to node 6. So, couldn't the query looks like [EN[4], ST[6]]? Sorry, I am a little bit confused.
•  » » 16 months ago, # ^ |   +5 In case 1, if $u$ is the ancestor then range $EN[u],ST[v]$ won't work because when you exit node $u$ you won't be able to reach node $v$ which is below $u$
 » 14 months ago, # |   0 Nice blog
 » 6 months ago, # | ← Rev. 2 →   0 I believe the range ${[ST(u)+1, EN(v)-1] \cup [ST(P), ST(P)]}$works regardless of whether $P=u$ holds, assuming $ST(u) \le ST(v)$.Worked example: 245030038