Please, try EDU on Codeforces! New educational section with videos, subtitles, texts, and problems. ×

### dacin21's blog

By dacin21, history, 2 years ago, , This blog assumes the reader is familiar with the basic concept of rolling hashes. There are some math-heavy parts, but one can get most of the ideas without understanding every detail.

The main focus of this blog is on how to choose the rolling-hash parameters to avoid getting hacked and on how to hack codes with poorly chosen parameters.

# Designing hard-to-hack rolling hashes

## Recap on rolling hashes and collisions

Recall that a rolling hash has two parameters (p, a) where p is the modulo and 0 ≤ a < p the base. (We'll see that p should be a big prime and a larger than the size of the alphabet.) The hash value of a string S = s 0... s n - 1 is given by For now, lets consider the simple problem of: given two strings S, T of equal length, decide whether they're equal by comparing their hash values h(S), h(T). Our algorithm declares S and T to be equal iff h(S) = h(T). Most rolling hash solutions are built on multiple calls to this subproblem or rely on the correctness of such calls.

Let's call two strings S, T of equal length with S ≠ T and h(S) = h(T) an equal-length collision. We want to avoid equal-length collisions, as they cause our algorithm to incorrectly assesses S and T as equal. (Note that our algorithms never incorrectly assesses strings a different.) For fixed parameters and reasonably small length, there are many more strings than possible hash values, so there always are equal-length collisions. Hence you might think that, for any rolling hash, there are inputs for which it is guaranteed to fail.

Luckily, randomization comes to the rescue. Our algorithm does not have to fix (p, a), it can randomly pick then according to some scheme instead. A scheme is reliable if we can prove that for arbitrary two string S, T, S ≠ T the scheme picks (p, a) such that h(S) ≠ h(T) with high probability. Note that the probability space only includes the random choices done inside the scheme; the input (S, T) is arbitrary, fixed and not necessarily random. (If you think of the input coming from a hack, then this means that no matter what the input is, our solution will not fail with high probability.)

I'll show you two reliable schemes. (Note that just because a scheme is reliable does not mean that your implementation is good. Some care has to be taken with the random number generator that is used.)

## Randomizing base

##### This part is based on a blog by rng_58. His post covers a more general hashing problem and is worth checking out.

This scheme uses a fixed prime p (i.e. 109 + 7 or 4·109 + 7) and picks a uniformly at random from . Let A be a random variable for the choice of a.

To prove that this scheme is good, consider two strings (S, T) of equal length and do some calculations Note that the left-hand side, let's call it P(A), is a polynomial of degree  ≤ n - 1 in A. P is non-zero as S ≠ T. The calculations show that h(S) = h(T) if and only if A is a root of P(A).

As p is prime and we are doing computations , we are working in a field. Over a field, any polynomial of degree  ≤ n - 1 has at most n - 1 roots. Hence there are at most n - 1 choices of a that lead to h(S) = h(T). Therefore So for any two strings (S, T) of equal length, the probability that they form an equal-length collision is at most . This is around 10 - 4 for n = 105, p = 109 + 7. Picking larger primes such as 231 - 1 or 4·109 + 7 can improve this a bit, but one needs more care with overflows.

### Tightness of bound

For now, this part only applies to primes with smooth p - 1, so it doesn't work for p = 109 + 7 for example. It would be interesting to find a construction that is computable and works in the general case.

The bound for this scheme is actually tight if . Consider and with

P(A) = A n - 1 - 1

As p is prime, is cyclic of order p - 1, hence there is a subgroup of order n - 1. Any then satisfies g n - 1 = 1, so P(A) has n - 1 distinct roots.

## Randomizing modulo

This scheme fixes a base and a bound N > a and picks a prime p uniformly at random from [N, 2N - 1].

To prove that this scheme is good, again, consider two strings (S, T) of equal length and do some calculations As , . As we chose a large enough, X ≠ 0. Moreover . An upper bound for the number of distinct prime divisors of X in [N, 2N - 1] is given by . By the prime density theorem, there are around primes in [N, 2N - 1]. Therefore Note that this bound is slightly worse than the one for randomizing the base. It is around 3·10 - 4 for n = 105, a = 26, N = 109.

## How to randomize properly

##### The following are good ways of initializing your random number generator.
• high precision time.
    chrono::duration_cast<chrono::nanoseconds>(chrono::high_resolution_clock::now().time_since_epoch()).count();


Either of the two should be fine. (In theory, high_resolution_clock should be better, but it somehow has lower precision than steady_clock on codeforces??)

• processor cycle counter
    __builtin_ia32_rdtsc();

• some heap address converted to an integer
    (uintptr_t) make_unique<char>().get();

    // pragma version
#pragma GCC target ("rdrnd")
uint32_t rdrand32(){
uint32_t ret;
assert(__builtin_ia32_rdrand32_step (&ret));
return ret;
}

    // asm version
uint32_t rd() {
uint32_t ret;
asm volatile("rdrand %0" :"=a"(ret) ::"cc");
return ret;
}


If you use a C++11-style rng (you should), you can use a combination of the above

    seed_seq seq{
(uint64_t) chrono::duration_cast<chrono::nanoseconds>(chrono::high_resolution_clock::now().time_since_epoch()).count(),
(uint64_t) __builtin_ia32_rdtsc(),
(uint64_t) (uintptr_t) make_unique<char>().get()
};
mt19937 rng(seq);
int base = uniform_int_distribution<int>(0, p-1)(rng);


Note that this does internally discard the upper 32 bits from the arguments and that this doesn't really matter, as the lower bits are harder to predict (especially in the first case with chrono.).

## Extension to multiple hashes

We can use multiple hashes (Even with the same scheme and same fixed parameters) and the hashes are independent so long as the random samples are independent. If the single hashes each fail with probability at most α1, ..., α k, the probability that all hashes fail is at most .

For example, if we use two hashes with p = 109 + 7 and randomized base, the probability of a collision is at most 10 - 8; for four hashes it is at most 10 - 16. Here the constants from slightly larger primes are more significant, for p = 231 - 1 the probabilities are around 2.1·10 - 9 and 4.7·10 - 18.

### Larger modulo

Using larger (i.e. 60 bit) primes would make collision less likely and not suffer from the accumulated factors of n in the error bounds. However, the computation of the rolling hash gets slower and more difficult, as there is no __int128 on codeforces.

One exception to this is the Mersenne prime p = 261 - 1; we can reduce by using bitshifts instead. (Thanks dmkozyrev for suggesting this.) The following code computes without __int128 and is only around 5 % slower than a 30 bit hash with modulo.

Code

A smaller factor can be gained by using unsigned types and p = 4·109 + 7.

Note that p = 264 (overflow of unsigned long long) is not prime and can be hacked regardless of randomization (see below).

## Extension to multiple comparisons

Usually, rolling hashes are used in more than a single comparison. If we rely on m comparison and the probability that a single comparison fails is p then the probability that any of the fail is at most m·p by a union bound. Note that when m = 105, we need at least two or three hashes for this to be small.

One has to be quite careful when estimating the number comparison we need to succeed. If we sort the hashes or put them into a set, we need to have pair-wise distinct hashes, so for n string a total of comparisons have to succeed. If n = 3·105, m ≈ 4.5·109, so we need three or four hashes (or only two if we use p = 261 - 1).

## Extension to strings of different length

If we deal with strings of different length, we can avoid comparing them by storing the length along the hash. This is not necessarily however, if we assume that no character hashes to 0. In that case, we can simple imagine we prepend the shorter strings with null-bytes to get strings of equal length without changing the hash values. Then the theory above applies just fine. (If some character (i.e. 'a') hashes to 0, we might produce strings that look the same but aren't the same in the prepending process (i.e. 'a' and 'aa').)

# Computing anti-hash tests

This section cover some technique that take advantage of common mistakes in rolling hash implementations and can mainly be used for hacking other solutions. Here's a table with a short summary of the methods.

Name Use case Runtime String length Notes
Thue-Morse Hash with overflow Θ(1) 210 Works for all bases simultaneously.
Birthday Small modulo  Can find multiple collisions.
Tree Large modulo  faster; longer strings
Multi-tree Large modulo  slower; shorter strings
Lattice reduction Medium-large alphabet, Multiple hashes  Great results for , good against multiple hashes. Bad on binary alphabet.
Composition Multiple hashes Sum of single runtimes Product of single string lengths Combines two attacks.

## Single hashes

### Thue–Morse sequence: Hashing with unsigned overflow ( p = 264, q arbitrary)

One anti-hash test that works for any base is the Thue–Morse sequence, generated by the following code.

code

See this blog for a detailed discussion. Note that the bound on the linked blog can be improved slightly, as X 2 - 1 is always divisible by 8 for odd X. (So we can use Q = 10 instead of Q = 11.)

### Birthday-attack: Hashing with 32-bit prime and fixed base ( p < 232 fixed, q fixed)

Hashes with a single small prime can be attacked via the birthday paradox. Fix a length l, let and pick k strings of length l uniformly at random. If l is not to small, the resulting hash values will approximately be uniformly distributed. By the birthday paradox, the probability that all of our picked strings hash to different values is Hence with probability we found two strings hashing to the same value. By repeating this, we can find an equal-length collision with high probability in . In practice, the resulting strings can be quite small (length  ≈ 6 for p = 109 + 7, not sure how to upper-bound this.).

More generally, we can compute m strings with equal hash value in using the same technique with .

### Tree-attack: Hashing with larger prime and fixed base ( p fixed, q fixed)

##### Thanks Kaban-5 and pavel.savchenkov for the link to some Russian comments describing this idea.

For large primes, the birthday-attack is to slow. Recall that for two strings (S, T) of equal length where α i = S i - T i satisfies . The tree-attack tries to find such that The attack maintains clusters C 1, ..., C k of coefficients. The sum S(C) of a cluster C is given by We can merge two clusters C 1 and C 2 to a cluster C 3 of sum S(C 1) - S(C 2) by multiplying all the α i from C 2 with  - 1 and joining the set of coefficients of C 1 and C 2. This operation can be implemented in constant time by storing the clusters as binary trees where each node stores its sum; the merge operation then adds a new node for C 3 with children C 1 and C 2 and sum S(C 1) - S(C 2). To ensure that the S(C 3) ≥ 0, swap C 1 and C 2 if necessary. The values of the α i are not explicitly stored, but they can be recomputed in the end by traversing the tree.

Initially, we start with n = 2 k and each α i = 1 in its own cluster. In a phase, we first sort the clusters by their sum and then merge adjacent pairs of clusters. If we encounter a cluster of sum 0 at any point, we finish by setting all α j not in that cluster to 0. If we haven't finished after k phases, try again with a bigger value of k.

For which values of k can we expect this to work? If we assume that the sums are initially uniformly distributed in , the maximum sum should decrease by a factor in phase i. After k phases, the maximum sum is around , so works. This produces strings of length in time. (A more formal analysis can be found in the paper 'Solving Medium-Density Subset Sum Problems in Expected Polynomial Time', section 2.2. The problem and algorithms in the paper are slightly different, but the approach similar.)

#### Multi-tree-attack

While the tree-attacks runs really fast, the resulting strings can get a little long. ( n = 2048 for p = 261 - 1.) We can spend more runtime to search for a shorter collision by storing the smallest m sums we can get in each cluster. (The single-tree-attack just uses m = 1.) Merging two clusters can be done in with a min-heap and a 2m-pointer walk. In order to get to m strings ASAP, we allow all values and exclude the trivial case where all α i are zero.

Analysing the expected value of k for this to work is quite difficult. Under the optimistic assumption that we reach m sums per node after steps, that the sums decrease as in the single tree attack and that we can expected a collision when they get smaller than m 2 by the birthday-paradox, we get . (A more realistic bound would be , which might be gotten by accounting for the birthday-paradox in the bound proven in the paper 'On Random High Density Subset Sums', Theorem 3.1.)

In practice, we can use m ≈ 105 to find a collision of length 128 for , p = 261 - 1 in around 0.4 seconds.

### Lattice-reduction attack: Single or multiple hashes over not-to-small alphabet

##### Thanks to hellman_ for mentioning this, check out his write-up on this topic here. There's also a write-up by someone else here.

As in the tree attack, we're looking for such that The set forms a lattice (A free -module embedded in a subspace of .) We're looking for an element in the lattice such that β = 0 and . We can penalize non-zero values of β by considering instead, then we seek to minimize . Unfortunately, this optimization problem is quite hard, so we try to minimize

α0 2 + ... + α n - 1 2 + β2

instead. The resulting problem is still hard, possibly NP-complete, but there are some good approximation algorithms available.

Similar to a vector space, we can define a basis in a lattice. For our case, a basis is given by A lattice reduction algorithm takes this basis and transforms it (by invertible matrices with determinant  ± 1) into another basis with approximately shortest vectors. Implementing them is quite hard (and suffers from precision errors or bignum slowdown), so I decided to use the builtin implementation in sage.

code

Sage offers two algorithms: LLL and BKZ, the former is faster but produces worse approximations, especially for longer strings. Analyzing them is difficult, so I experimented a bit by fixing , p = 261 - 1 and fixing a 1, ..., a n randomly and searching for a short anti-hash test with both algorithms. The results turned out really well.

Experiment data

Note that this attack does not work well for small (i.e binary) alphabets when n > 1 and that the characters have to hash to consecutive values, so this has to be the first attack if used in a composition attack.

## Composition-attack: Multiple hashes

##### Credit for this part goes to ifsmirnov, I found this technique in his jngen library.

Using two or more hashes is usually sufficient to protect from a direct birthday-attack. For two primes, there are N = p 1·p 2 possible hash values. The birthday-attack runs in , which is  ≈ 1010 for primes around 109. Moreover, the memory usage is more than bytes (If you only store the hashes and the rng-seed), which is around 9.5 GB.

The key idea used to break multiple hashes is to break them one-by-one.

• First find an equal-length collision (by birthday-attack) for the first hash h 1, i.e. two strings S, T, S ≠ T of equal length with h 1(S) = h 1(T). Note that strings of equal length built over the alphabet S, T (i.e. by concatenation of some copies of S with some copies of T and vice-versa) will now hash to the same value under h 1.
• Then use S and T as the alphabet when searching for an equal-length collision (by birthday-attack again) for the second hash h 2. The result will automatically be a collision for h 1 as well, as we used S, T as the alphabet.

This reduces the runtime . Note that this also works for combinations of a 30-bit prime hash and a hash mod 264 if we use the Thue–Morse sequence in place of the second birthday attack. Similarly, we can use tree- instead of birthday-attacks for larger modulos.

Another thing to note is that string length grows rapidly in the number of hashes. (Around , the alphabet size is reduced to 2 after the first birthday-attack. The first iteration has a factor of 2 in practice.) If we search for more than 2 strings with equal hash value in the intermediate steps, the alphabet size will be bigger, leading to shorter strings, but the runtime of the birthday-attacks gets slower ( for 3 strings, for example.).

On codeforces, quite a lot of people randomize their hashes. (Un-)Fortunately, many of them do it an a suboptimal way. This section covers some of the ways people screw up their hash randomizations and ways to hack their code.

This section applies more generally to any type of randomized algorithm in an environment where other participants can hack your solutions.

### Fixed seed

If the seed of the rng is fixed, it always produces the same sequence of random numbers. You can just run the code to see which numbers get randomly generated and then find an anti-hash test for those numbers.

### Picking from a small pool of bases (rand() % 100)

Note that rand() % 100 produced at most 100 distinct values (0, ..., 99). We can just find a separate anti-hash test for every one of them and then combine the tests into a single one. (The way your combine tests is problem-specific, but it works for most of the problems.)

### More issues with rand()

On codeforces, rand() produces only 15-bit values, so at most 215 different values. While it may take a while to run 215 birthday-attacks (estimated 111 minutes for p = 109 + 7 using a single thread on my laptop), this can cause some big issues with some other randomized algorithms.

Edit: This type of hack might be feasible if we use multi-tree-attacks. For , running 215 multi-tree attacks with m = 104 takes around 2 minutes and produces an output of 5.2·105 characters. This is still slightly to large for most problems, but could be split up into multiple hacks in an open hacking phase, for example.

In C++11 you can use mt19937 and uniform_int_distribution instead of rand().

### Low-precision time (Time(NULL))

Time(NULL) only changes once per second. This can be exploited as follows

1. Pick a timespan Δ T.
2. Find an upper bound T for the time you'll need to generate your tests.
3. Figure out the current value T 0 of Time(NULL) via custom invocation.
4. For t = 0, ..., (Δ T) - 1, replace Time(NULL) with T 0 + T + t and generate an anti-test for this fixed seed.
5. Submit the hack at time T 0 + T.

If your hack gets executed within the next Δ T seconds, Time(NULL) will be a value for which you generated an anti-test, so the solution will fail.

### Random device on MinGW (std::random_device)

Note that on codeforces specifically, std::random_device is deterministic and will produce the same sequence of numbers. Solutions using it can be hacked just like fixed seed solutions.

# Notes

• If I made a mistake or a typo in a calculation, or something is unclear, please comment.
• If you have your own hash randomization scheme, way of seeding the rng or anti-hash algorithm that you want to discuss, feel free to comment on it below.
• I was inspired to write this blog after the open hacking phase of round #494 (problem F). During (and after) the hacking phase I figured out how to hack many solution that I didn't know how to hack beforehand. I (un-)fortunately had to go to bed a few hours in (my timezone is UTC + 2), so quite a few hackable solutions passed. math, Comments (26)
 » Auto comment: topic has been updated by dacin21 (previous revision, new revision, compare).
 » 2 years ago, # | ← Rev. 6 →   Suppose that we use some fixed prime p (for example, p = 109 + 7, the reason to consider this number is that p - 1 = 2q, where q = 5·108 + 3 is prime, has only one small prime factor: 2) and randomize the base. This clearly rules out your high collision probability example because it only can happen when n = 3 or n ≥ q + 1, neither of this situation is close to the model case when n ≈ 105. Can we find a bad case in this situation?I have a proof that there are two strings that have same hash with probability at least , where is some constant. However, I don't know any way to construct such two strings in reasonable time. Does anybody know a way (possibly completely different)? Proof, some math hereConsider — the field of remainders modulo p and the ring of polynomials over it. Clearly, finding a polynomial in of degree less than n with coefficients in range [ - (|Σ| — 1), (|Σ| — 1)] with a lot (respectively to n, of course) of roots (in ) is sufficient for finding a hash collision. Let k:  = |Σ| - 1) and consider a polynomial (we can take any r distinct elements of as roots of q, not necessarily 0, ..., r - 1), where r is some integer that we will choose later. Consider all polynomials from with degree less than n and coefficients in range [0, k]. There are kn of them. If k n > p r, then two of them have the same remainder modulo q, because possible remainders modulo q are polynomials from our ring with degree  < r; their difference is the polynomial P we were looking for — it is divisible by q, therefore it has r roots, it has degree less than n and its coefficients are from range [ - k, k]. If we choose , then k n > p r holds, therefore we found a polynomial of degree n with at least cn roots, where (probably up to  ± 1 to some direction). In the case k = 26 - 1 = 25, p ≈ 109 we have c ≈ 0.16, which is large enough I guess.Moreover, one may heuristically apply birthday paradox argument here, making c even bigger. This does not solve the problem of actually finding polynomials with same remainders modulo q, though.P. S. It is also possible to break single hash modulo 64-bit (link to the problem, it is in Russian though, Breaking hashing). I hope Gassa wouldn't mind explaining the solution, he will definitely do this better than me. In fact, I remember somebody claiming that they can break fixed hashes modulo number of order 10100.
•  » »
•  » » » Big thanks to both of you for linking me this, I'll add a subsection on this. (The attack can be improved by storing multiple values at each node; if we store 104 values per node, the resulting length drops to 64 for p ≈ 231 and 256 for p ≈ 261 and with 105 values it drops to 32 and 128 and runs in  < 1 s.)
 » 2 years ago, # | ← Rev. 6 →   Thank you very much for your post! By the way, today I came to the conclusion, that we can calculate double rolling hash without operations %, if we use two modules: prime m1 = 2^31-1 and not prime m2 = 2^64. For hash by m2 we just need to work in unsigned long long.Let's find out how to take the remainder of the division by m1. A simple analogy with the 10-base number systemIf we have number in 10-base system: number = digit + digit * 10 + digit * 100 + digit * 1000 + ...And want calculate remainder from division by mod 9, we just need to accumulate all digits and repeat while sum >= 9, because 10^p % 9 == 1number % 9 = (digit + digit + digit + ...) % 9This works for 9 * k not as you expect. This method gives the result from 1 to 9 for all positive values and 0 only for 0Let x be from 1 to (m1-1)^2 — result from multiplication of two remainders. For getting remainder modulo 2^31-1 we need to set x a sum all digits in base 2^31 and repeat while x >= 2^31-1. x = (x >> 31) + (x & 2147483647); x = (x >> 31) + (x & 2147483647); return x; But it works for (2^31-1) * k numbers not as you expect. This method gives the result from 1 to 2^31-1 for all positive values and 0 only for 0. x % (2^31-1) is number from [0..2^31-2]This method definitely gives number from [1..2^31-1].It works great for rolling hashes if we cut 0 and if we compare results from one type of method (works only with remainders [1..2^31-1] against [0..2^31-2]). Maybe it can speed up rolling hashes solutions.
•  » » Thanks a lot for your comment.First of all, don't use 264 as your modulo! It can be hacked regardless of the base. (See the section 'Hashing with unsigned overflow ( p = 264, q arbitrary)'). Combining it with another hash does not save you, see the section 'Multiple hashes', the Thue–Morse sequence can be used there as well.Second of, your bit-trick technique can be used to map to as well, just add 1 to x beforehand and subtract 1 from x afterwards. It is probably more useful if we use p = 261 - 1 (with some more bit-tricks), as the alternative with __int128 does not exists on codeforces. Badly tested code for that pconst uint64_t mod = (1ull<<61) - 1; uint64_t modmul(uint64_t a, uint64_t b){ uint64_t l1 = (uint32_t)a, h1 = a>>32, l2 = (uint32_t)b, h2 = b>>32; uint64_t l = l1*l2, m = l1*h2 + l2*h1, h = h1*h2; uint64_t ret = (l&mod) + (l>>61) + (h << 3) + (m >> 29) + (m << 35 >> 3) + 1; ret = (ret & mod) + (ret>>61); ret = (ret & mod) + (ret>>61); return ret-1; } Basic idea: Split the number into lower 32 and upper 29 bits. Then multiply the parts The lowest part is properly shifted, but it might be up to 264 - 233 + 1, so we do one reduction on it. The highest part needs to be left-shifted by 64; this is equivalent to shifting by 3. (The result is at most (229 - 1)2·8 < 261.) The middle part has to be left-shifted by 32. This would lead to overflow, so we right shift the higher bits by 29 instead. Now we can add everything without overflow and reduce. I'll add this (with p = 261 - 1) to the section about larger modulos once I get around to test and benchmark it.
•  » » » 2 years ago, # ^ | ← Rev. 2 →   Thanks, please tell me how fast hacker can hack my solution if he knows: I'm using double polynomial hash with p1 = 2^31-1 and p2 = 2^64 I'm using random generation of odd q with uniform distribution from (256, 2^31-1) with std::random_device, std::mt19937, std::uniform_int_distribution For 2^64 exist anti-hash test regardless of the base q Time of running solution on server in seconds, for examplt 14:52:31.
•  » » » » At first I misread std::random_device as high_resolution_clock and it seemed quite hard. (The main issue is to break p 1 with less than characters.) But on codeforces, the former always returns 3499211612 as the first number. I can hack you in less than a minute as follows Copy/retype your random-generation code and run it in a custom invocation to get your q. (As noted above, it will always be the same.) Run a birthday attack against p 1 (takes  < 1 second). Build a Thue–Morse sequence using the strings I got from the birthday-attack instead of 'A' and 'B'.
•  » » » » » 2 years ago, # ^ | ← Rev. 2 →   You need to combine sqrt(p1) ~=~ 2^16 chars with 2^11 of Thue–Morse string, this is 2^27 = 134.217.728 chars, if I understand correctly, it is possible? I never seen input greater than 5 * 10^6 chars.I asked it because I wrote tutorial for beginners last day and in this tutorial I used two modules: 2^64, 10^9+123 and random point, generated with std::mt19937, std::chrono::high_resolution_clock and std::uniform_int_distribution and wrote that using std::time(0) or std::random_device in MinGW not safety, but now I understand that you need very big input to hack solution with fixed point and maybe don't need randomization?
•  » » » » » » You're mixing apples and oranges. For the birthday attack, we need to generate strings. These strings can be quite short, in practice around 6 characters each. We then pick two strings that hash to same value and use them in the Thue-Morse sequence. So the total length would be 6·2·210 = 12288. (210 sufficies, 211 was based of a non-tight bound from the linked post on hashes mod 264.)If you use 264 and a prime around 109 with good randomization, I haven't figured out a good hack yet. (Note that this does not mean you can't be hacked, someone might find a better attack.) The best I thought of was building a generalized Thue-Morse sequence in base , so that all pairs collide mod 264 and there is a good change for a birthday-collision mod p. This would lead to massive input however (Estimated 210·6·45'000 ≈ 3·108), as it involves submitting an input with long strings. So I guess there's a good chance you won't be hacked, but I myself prefer something where I can proof that it's hard for me to get hacked and not rely on other not figuring out some new technique.
•  » » » » » » » Now I understood, thanks
•  » » » 2 years ago, # ^ | ← Rev. 3 →   I not understood why this method not works on this problem on generated test. Code on ideone.com or my solution is bad.I generated test with length 5000: s = 'v' * 2500 + '-' * 2500 t = '~' * 2500 + 'v' * 2500 True answer is 2500, but it gets 2499 or 2500 (Answer depends on launch with random point).UPD: Sorry, bug in function diff, my bad:a and b unsigned long long and it is not correct (compare with zero) in function sub:return (a -= b) < 0 ? a + mod : a;It should be like this:return (a -= b) >= mod ? a + mod : a;
 » My randomization function: uint32_t rd() { uint32_t res; #ifdef __MINGW32__ asm volatile("rdrand %0" :"=a"(res) ::"cc"); #else res = std::random_device()(); #endif return res; } 
•  » » Thanks, added it and a version that uses the corresponding built-in function instead of asm.
 » 2 years ago, # | ← Rev. 3 →   Double hashing has never failed me. So I think you don't have to worry for collision in CE if you use double hashing. My primes for mod are 109 + 7 and 109 + 9 and I use two random primes like 737 or 3079 or 4001 as base.
•  » » If the input is random so that we can assume the hash-values are uniformly distributed, you're fine.But if you don't randomize your bases at runtime (and it sounds like you don't), someone might hack your solution. (And AFAIK it makes no difference whether the base is prime or not.)
•  » » » e-maxx has it mentioned that we should choose base as prime > max character in string. It says for eg. if string is all lowercase choose 31. But it didn't mention why.
•  » » Now that you are no more random :D, Get ready to... :D
 » Auto comment: topic has been updated by dacin21 (previous revision, new revision, compare).
 » On abusing multiple known modulos: you can quickly generate short hack-tests using lattice reduction methods (LLL). There was a related challenge at TokyoWesterns CTF (an infosec competition) and I made a writeup for that (or another writeup). There you have 8 32-bit modulos and the generated hack-test is less than 100 symbols. (The problem there is palindrome search but the method is generic).
•  » » Thank, lattice reduction works great for multiple hashes if the alphabet size is not to small. I added a section on it with some modified code. For many 61-bit hashes, the LLL algorithm deteriorates, but the BKZ algorithm in sage works instead.
 » Thanks for your awesome post !Just another suggestion or request: it would be great if you would provide a sample full code combining all of the ideas at the end of the post.
 » Awesome post. .simple and really informative..!!
 » 5 months ago, # | ← Rev. 4 →   Although it sounds rather weird to bump such an old blog, I want to point out that you can actually perform a birthday attack with $O(1)$ memory using Floyd's cycle finding algorithm (similar to pollard-rho).Assuming your hash function is $H: M\rightarrow D$, let's find a random mapping $F: D \rightarrow M$ that generates a random message using a hash value as seed. Assume $F$ is injective.Let $x:=1, y:=1$ in the beginning, $T(x)=H(F(x))$ and repeat $x:=T(T(x)), y:=T(y)$ until $x=y$. We've found a collision and we need to retrace how it happened. This can be done by resetting $x:=1$ and repeat $x:=T(x), y:=T(y)$ until $T(x)=T(y)$. $(F(x),F(y))$ will be such a collision pair. CorrectnessAssume you've understood the correctness of pollard-rho. Notice that by the end of the first phrase, $x$ is faster than $y$ by some multiple of cycle length, thus the round passed for $y$ is also some multiple of cycle length. Therefore, we can keep that advantage and trace from the beginning back again. Sample Implementation#pragma GCC optimize("-Ofast","-funroll-all-loops","-ffast-math") #include using namespace std; typedef unsigned long long ull; const ull MOD=100000000000009ULL; //runtime ~1s //const ull MOD=10000000000000009ULL; //runtime ~10s //const ull MOD=1000000000000000009ULL; //runtime ~4min inline ull my_hash(string s) { //safest hash ever, even contains overflow ull h=0; for(auto c:s) h=(h*23333+c)%MOD; return h; } string gen_str(ull u) { string s; u=u*819039758391039471ULL+283958671940506ULL; while(u) s.push_back(u%26+'a'),u/=26; return s; } //my_hash(gen_str(u)) inline ull next_hash(ull u) { ull h=0; u=u*819039758391039471ULL+283958671940506ULL; while(u) h=(h*23333+(char(u%26+'a')))%MOD,u/=26; return h; } int main() { ull sh=1,p,q,x=sh,y=sh; while(1) { p=x, x=next_hash(p); p=x, x=next_hash(p); q=y, y=next_hash(q); if(x==y) break; } if(y==sh) { puts("Failed. :("); return 0; } x=sh; while(x!=y) p=x, x=next_hash(p), q=y, y=next_hash(q); string a=gen_str(p),b=gen_str(q); cout<<"hash("< using namespace std; typedef unsigned long long ull; const ull MOD=100000000000009ULL; //runtime ~0.6s //const ull MOD=10000000000000009ULL; //runtime ~4.4s //const ull MOD=1000000000000000009ULL; //runtime ~2min //safest hash ever inline ull my_hash(string s) { ull h=0; for(auto c:s) h=(h*23333+c)%MOD; return h; } string gen_str(ull u) { string s; u=u*819039758391039471ULL+283958671940506ULL; while(u) s.push_back(u%26+'a'),u/=26; return s; } //my_hash(gen_str(u)) inline ull next_hash(ull u) { ull h=0; u=u*819039758391039471ULL+283958671940506ULL; while(u) h=(h*23333+(char(u%26+'a')))%MOD,u/=26; return h; } unordered_map> w; int main() { ull sh=1,x=sh,y,p,q; int c0,c1; while(1) { ull u=x; int c=0; do u=next_hash(u),++c; while((u-sh)%100007); if(w.count(u)) {c0=c; c1=w[u].first; y=w[u].second; break;} w[u]=make_pair(c,x); x=u; } int g=min(c0,c1); c0-=g; c1-=g; while(c0--) p=x,x=next_hash(x); while(c1--) q=y,y=next_hash(y); while(x!=y) p=x,x=next_hash(x), q=y,y=next_hash(y); string a=gen_str(p),b=gen_str(q); cout<<"hash("<
 » 4 months ago, # | ← Rev. 3 →   When using multiple hashes, is it necessary to use different modulo, or is it ok to just a different base?Update: It's ok just to change the base.
 » loved it