This blog assumes the reader is familiar with the basic concept of rolling hashes. There are some math-heavy parts, but one can get most of the ideas without understanding every detail.

The main focus of this blog is on how to choose the rolling-hash parameters to avoid getting hacked and on how to hack codes with poorly chosen parameters.

# Designing hard-to-hack rolling hashes

## Recap on rolling hashes and collisions

Recall that a rolling hash has two parameters (*p*, *a*) where *p* is the modulo and 0 ≤ *a* < *p* the base. (We'll see that *p* should be a big prime and *a* larger than the size of the alphabet.) The hash value of a string *S* = *s*_{0}... *s*_{n - 1} is given by

For now, lets consider the simple problem of: given two strings *S*, *T* of equal length, decide whether they're equal by comparing their hash values *h*(*S*), *h*(*T*). Our algorithm declares *S* and *T* to be equal iff *h*(*S*) = *h*(*T*). Most rolling hash solutions are built on multiple calls to this subproblem or rely on the correctness of such calls.

Let's call two strings *S*, *T* of equal length with *S* ≠ *T* and *h*(*S*) = *h*(*T*) an **equal-length collision**. We want to avoid equal-length collisions, as they cause our algorithm to incorrectly assesses *S* and *T* as equal. (Note that our algorithms never incorrectly assesses strings a different.) For fixed parameters and reasonably small length, there are many more strings than possible hash values, so there always are equal-length collisions. Hence you might think that, for any rolling hash, there are inputs for which it is guaranteed to fail.

Luckily, randomization comes to the rescue. Our algorithm does not have to fix (*p*, *a*), it can randomly pick then according to some scheme instead. A scheme is **reliable** if we can prove that for arbitrary two string *S*, *T*, *S* ≠ *T* the scheme picks (*p*, *a*) such that *h*(*S*) ≠ *h*(*T*) with high probability. Note that the probability space only includes the random choices done inside the scheme; the input (*S*, *T*) is arbitrary, fixed and not necessarily random. (If you think of the input coming from a hack, then this means that no matter what the input is, our solution will not fail with high probability.)

I'll show you two reliable schemes. (Note that just because a scheme is reliable does not mean that your implementation is good. Some care has to be taken with the random number generator that is used.)

## Randomizing base

##### This part is based on a blog by rng_58. His post covers a more general hashing problem and is worth checking out.

This scheme uses a fixed **prime** *p* (i.e. 10^{9} + 7 or 4·10^{9} + 7) and picks *a* uniformly at random from . Let *A* be a random variable for the choice of *a*.

To prove that this scheme is good, consider two strings (*S*, *T*) of equal length and do some calculations

Note that the left-hand side, let's call it *P*(*A*), is a polynomial of degree ≤ *n* - 1 in *A*. *P* is non-zero as *S* ≠ *T*. The calculations show that *h*(*S*) = *h*(*T*) if and only if *A* is a root of *P*(*A*).

As *p* is prime and we are doing computations , we are working in a field. Over a field, any polynomial of degree ≤ *n* - 1 has at most *n* - 1 roots. Hence there are at most *n* - 1 choices of *a* that lead to *h*(*S*) = *h*(*T*). Therefore

So for any two strings (*S*, *T*) of equal length, the probability that they form an equal-length collision is at most . This is around 10^{ - 4} for *n* = 10^{5}, *p* = 10^{9} + 7. Picking larger primes such as 2^{31} - 1 or 4·10^{9} + 7 can improve this a bit, but one needs more care with overflows.

### Tightness of bound

For now, this part only applies to primes with smooth *p* - 1, so it doesn't work for *p* = 10^{9} + 7 for example. It would be interesting to find a construction that is computable and works in the general case.

The bound for this scheme is actually tight if . Consider and with

*P*(

*A*) =

*A*

^{n - 1}- 1

As *p* is prime, is cyclic of order *p* - 1, hence there is a subgroup of order *n* - 1. Any then satisfies *g*^{n - 1} = 1, so *P*(*A*) has *n* - 1 distinct roots.

## Randomizing modulo

This scheme fixes a base and a bound *N* > *a* and picks a **prime** *p* uniformly at random from [*N*, 2*N* - 1].

To prove that this scheme is good, again, consider two strings (*S*, *T*) of equal length and do some calculations

As , . As we chose *a* large enough, *X* ≠ 0. Moreover . An upper bound for the number of distinct prime divisors of *X* in [*N*, 2*N* - 1] is given by . By the prime density theorem, there are around primes in [*N*, 2*N* - 1]. Therefore

Note that this bound is slightly worse than the one for randomizing the base. It is around 3·10^{ - 4} for *n* = 10^{5}, *a* = 26, *N* = 10^{9}.

## How to randomize properly

##### The following are good ways of initializing your random number generator.

- high precision time.

```
chrono::duration_cast<chrono::nanoseconds>(chrono::high_resolution_clock::now().time_since_epoch()).count();
chrono::duration_cast<chrono::nanoseconds>(chrono::steady_clock::now().time_since_epoch()).count();
```

Either of the two should be fine. (In theory, `high_resolution_clock`

should be better, but it somehow has lower precision than `steady_clock`

on codeforces??)

- processor cycle counter

```
__builtin_ia32_rdtsc();
```

- some heap address converted to an integer

```
(uintptr_t) make_unique<char>().get();
```

- processor randomness (needs either pragma or asm) (Thanks halyavin for suggesting this.)

```
// pragma version
#pragma GCC target ("rdrnd")
uint32_t rdrand32(){
uint32_t ret;
assert(__builtin_ia32_rdrand32_step (&ret));
return ret;
}
```

```
// asm version
uint32_t rd() {
uint32_t ret;
asm volatile("rdrand %0" :"=a"(ret) ::"cc");
return ret;
}
```

If you use a C++11-style rng (you should), you can use a combination of the above

```
seed_seq seq{
(uint64_t) chrono::duration_cast<chrono::nanoseconds>(chrono::high_resolution_clock::now().time_since_epoch()).count(),
(uint64_t) __builtin_ia32_rdtsc(),
(uint64_t) (uintptr_t) make_unique<char>().get()
};
mt19937 rng(seq);
int base = uniform_int_distribution<int>(0, p-1)(rng);
```

Note that this does internally discard the upper 32 bits from the arguments and that this doesn't really matter, as the lower bits are harder to predict (especially in the first case with chrono.).

##### See the section on 'Abusing bad randomization' for some bad examples.

## Extension to multiple hashes

We can use multiple hashes (Even with the same scheme and same fixed parameters) and the hashes are independent so long as the random samples are independent. If the single hashes each fail with probability at most α_{1}, ..., α_{k}, the probability that all hashes fail is at most .

For example, if we use two hashes with *p* = 10^{9} + 7 and randomized base, the probability of a collision is at most 10^{ - 8}; for four hashes it is at most 10^{ - 16}. Here the constants from slightly larger primes are more significant, for *p* = 2^{31} - 1 the probabilities are around 2.1·10^{ - 9} and 4.7·10^{ - 18}.

### Larger modulo

Using larger (i.e. 60 bit) primes would make collision less likely and not suffer from the accumulated factors of *n* in the error bounds. However, the computation of the rolling hash gets slower and more difficult, as there is no `__int128`

on codeforces.

One exception to this is the Mersenne prime *p* = 2^{61} - 1; we can reduce by using bitshifts instead. (Thanks dmkozyrev for suggesting this.) The following code computes without `__int128`

and is only around 5 % slower than a 30 bit hash with modulo.

**Code**

A smaller factor can be gained by using unsigned types and *p* = 4·10^{9} + 7.

Note that *p* = 2^{64} (overflow of unsigned long long) is **not prime** and can be **hacked** regardless of randomization (see below).

## Extension to multiple comparisons

Usually, rolling hashes are used in more than a single comparison. If we rely on *m* comparison and the probability that a single comparison fails is *p* then the probability that any of the fail is at most *m*·*p* by a union bound. Note that when *m* = 10^{5}, we need at least two or three hashes for this to be small.

One has to be quite careful when estimating the number comparison we need to succeed. If we sort the hashes or put them into a set, we need to have pair-wise distinct hashes, so for *n* string a total of comparisons have to succeed. If *n* = 3·10^{5}, *m* ≈ 4.5·10^{9}, so we need three or four hashes (or only two if we use *p* = 2^{61} - 1).

## Extension to strings of different length

If we deal with strings of different length, we can avoid comparing them by storing the length along the hash. This is not necessarily however, if we assume that **no character hashes to** 0. In that case, we can simple imagine we prepend the shorter strings with null-bytes to get strings of equal length without changing the hash values. Then the theory above applies just fine. (If some character (i.e. 'a') hashes to 0, we might produce strings that look the same but aren't the same in the prepending process (i.e. 'a' and 'aa').)

# Computing anti-hash tests

This section cover some technique that take advantage of common mistakes in rolling hash implementations and can mainly be used for hacking other solutions. Here's a table with a short summary of the methods.

Name | Use case | Runtime | String length | Notes |
---|---|---|---|---|

Thue-Morse | Hash with overflow | Θ(1) | 2^{10} | Works for all bases simultaneously. |

Birthday | Small modulo | Can find multiple collisions. | ||

Tree | Large modulo | faster; longer strings | ||

Multi-tree | Large modulo | slower; shorter strings | ||

Lattice reduction | Medium-large alphabet, Multiple hashes | Great results for , good against multiple hashes. Bad on binary alphabet. | ||

Composition | Multiple hashes | Sum of single runtimes | Product of single string lengths | Combines two attacks. |

## Single hashes

### Thue–Morse sequence: Hashing with unsigned overflow (*p* = 2^{64}, *q* arbitrary)

One anti-hash test that works for *any* base is the Thue–Morse sequence, generated by the following code.

**code**

See this blog for a detailed discussion. Note that the bound on the linked blog can be improved slightly, as *X*^{2} - 1 is always divisible by 8 for odd *X*. (So we can use *Q* = 10 instead of *Q* = 11.)

### Birthday-attack: Hashing with 32-bit prime and fixed base (*p* < 2^{32} fixed, *q* fixed)

Hashes with a single small prime can be attacked via the birthday paradox. Fix a length *l*, let and pick *k* strings of length *l* uniformly at random. If *l* is not to small, the resulting hash values will approximately be uniformly distributed. By the birthday paradox, the probability that all of our picked strings hash to different values is

Hence with probability we found two strings hashing to the same value. By repeating this, we can find an equal-length collision with high probability in . In practice, the resulting strings can be quite small (length ≈ 6 for *p* = 10^{9} + 7, not sure how to upper-bound this.).

More generally, we can compute *m* strings with equal hash value in using the same technique with .

### Tree-attack: Hashing with larger prime and fixed base (*p* fixed, *q* fixed)

##### Thanks Kaban-5 and pavel.savchenkov for the link to some Russian comments describing this idea.

For large primes, the birthday-attack is to slow. Recall that for two strings (*S*, *T*) of equal length

where α_{i} = *S*_{i} - *T*_{i} satisfies . The tree-attack tries to find such that

The attack maintains clusters *C*_{1}, ..., *C*_{k} of coefficients. The **sum** *S*(*C*) of a cluster *C* is given by

We can merge two clusters *C*_{1} and *C*_{2} to a cluster *C*_{3} of sum *S*(*C*_{1}) - *S*(*C*_{2}) by multiplying all the α_{i} from *C*_{2} with - 1 and joining the set of coefficients of *C*_{1} and *C*_{2}. This operation can be implemented in constant time by storing the clusters as binary trees where each node stores its sum; the merge operation then adds a new node for *C*_{3} with children *C*_{1} and *C*_{2} and sum *S*(*C*_{1}) - *S*(*C*_{2}). To ensure that the *S*(*C*_{3}) ≥ 0, swap *C*_{1} and *C*_{2} if necessary. The values of the α_{i} are not explicitly stored, but they can be recomputed in the end by traversing the tree.

Initially, we start with *n* = 2^{k} and each α_{i} = 1 in its own cluster. In a phase, we first sort the clusters by their sum and then merge adjacent pairs of clusters. If we encounter a cluster of sum 0 at any point, we finish by setting all α_{j} not in that cluster to 0. If we haven't finished after *k* phases, try again with a bigger value of *k*.

For which values of *k* can we expect this to work? If we assume that the sums are initially uniformly distributed in , the maximum sum should decrease by a factor in phase *i*. After *k* phases, the maximum sum is around , so works. This produces strings of length in time. (A more formal analysis can be found in the paper 'Solving Medium-Density Subset Sum Problems in Expected Polynomial Time', section 2.2. The problem and algorithms in the paper are slightly different, but the approach similar.)

#### Multi-tree-attack

While the tree-attacks runs really fast, the resulting strings can get a little long. (*n* = 2048 for *p* = 2^{61} - 1.) We can spend more runtime to search for a shorter collision by storing the smallest *m* sums we can get in each cluster. (The single-tree-attack just uses *m* = 1.) Merging two clusters can be done in with a min-heap and a 2*m*-pointer walk. In order to get to *m* strings ASAP, we allow all values and exclude the trivial case where all α_{i} are zero.

Analysing the expected value of *k* for this to work is quite difficult. Under the optimistic assumption that we reach *m* sums per node after steps, that the sums decrease as in the single tree attack and that we can expected a collision when they get smaller than *m*^{2} by the birthday-paradox, we get . (A more realistic bound would be , which might be gotten by accounting for the birthday-paradox in the bound proven in the paper 'On Random High Density Subset Sums', Theorem 3.1.)

In practice, we can use *m* ≈ 10^{5} to find a collision of length 128 for , *p* = 2^{61} - 1 in around 0.4 seconds.

### Lattice-reduction attack: Single or multiple hashes over not-to-small alphabet

##### Thanks to hellman_ for mentioning this, check out his write-up on this topic here. There's also a write-up by someone else here.

As in the tree attack, we're looking for such that

The set

forms a **lattice** (A free -module embedded in a subspace of .) We're looking for an element in the lattice such that β = 0 and . We can penalize non-zero values of β by considering

instead, then we seek to minimize . Unfortunately, this optimization problem is quite hard, so we try to minimize

_{0}

^{2}+ ... + α

_{n - 1}

^{2}+ β

^{2}

instead. The resulting problem is still hard, possibly NP-complete, but there are some good approximation algorithms available.

Similar to a vector space, we can define a basis in a lattice. For our case, a basis is given by

A lattice reduction algorithm takes this basis and transforms it (by invertible matrices with determinant ± 1) into another basis with approximately shortest vectors. Implementing them is quite hard (and suffers from precision errors or bignum slowdown), so I decided to use the builtin implementation in sage.

**code**

Sage offers two algorithms: `LLL`

and `BKZ`

, the former is faster but produces worse approximations, especially for longer strings. Analyzing them is difficult, so I experimented a bit by fixing , *p* = 2^{61} - 1 and fixing *a*_{1}, ..., *a*_{n} randomly and searching for a short anti-hash test with both algorithms. The results turned out really well.

**Experiment data**

Note that this attack does not work well for small (i.e binary) alphabets when *n* > 1 and that the characters have to **hash to consecutive** values, so this has to be the first attack if used in a composition attack.

## Composition-attack: Multiple hashes

##### Credit for this part goes to ifsmirnov, I found this technique in his jngen library.

Using two or more hashes is usually sufficient to protect from a direct birthday-attack. For two primes, there are *N* = *p*_{1}·*p*_{2} possible hash values. The birthday-attack runs in , which is ≈ 10^{10} for primes around 10^{9}. Moreover, the memory usage is more than bytes (If you only store the hashes and the rng-seed), which is around 9.5 GB.

The key idea used to break multiple hashes is to break them one-by-one.

- First find an equal-length collision (by birthday-attack) for the first hash
*h*_{1}, i.e. two strings*S*,*T*,*S*≠*T*of equal length with*h*_{1}(*S*) =*h*_{1}(*T*). Note that strings of equal length built over the alphabet*S*,*T*(i.e. by concatenation of some copies of*S*with some copies of*T*and vice-versa) will now hash to the same value under*h*_{1}. - Then use
*S*and*T*as the alphabet when searching for an equal-length collision (by birthday-attack again) for the second hash*h*_{2}. The result will automatically be a collision for*h*_{1}as well, as we used*S*,*T*as the alphabet.

This reduces the runtime . Note that this also works for combinations of a 30-bit prime hash and a hash mod 2^{64} if we use the Thue–Morse sequence in place of the second birthday attack. Similarly, we can use tree- instead of birthday-attacks for larger modulos.

Another thing to note is that string length grows rapidly in the number of hashes. (Around , the alphabet size is reduced to 2 after the first birthday-attack. The first iteration has a factor of 2 in practice.) If we search for more than 2 strings with equal hash value in the intermediate steps, the alphabet size will be bigger, leading to shorter strings, but the runtime of the birthday-attacks gets slower ( for 3 strings, for example.).

## Abusing bad randomization

On codeforces, quite a lot of people randomize their hashes. (Un-)Fortunately, many of them do it an a suboptimal way. This section covers some of the ways people screw up their hash randomizations and ways to hack their code.

This section applies more generally to any type of randomized algorithm in an environment where other participants can hack your solutions.

### Fixed seed

If the seed of the rng is fixed, it always produces the same sequence of random numbers. You can just run the code to see which numbers get randomly generated and then find an anti-hash test for those numbers.

### Picking from a small pool of bases (`rand() % 100`

)

Note that `rand() % 100`

produced at most 100 distinct values (0, ..., 99). We can just find a separate anti-hash test for every one of them and then combine the tests into a single one. (The way your combine tests is problem-specific, but it works for most of the problems.)

### More issues with `rand()`

On codeforces, `rand()`

produces only 15-bit values, so at most 2^{15} different values. While it may take a while to run 2^{15} birthday-attacks (estimated 111 minutes for *p* = 10^{9} + 7 using a single thread on my laptop), this can cause some big issues with some other randomized algorithms.

Edit: This type of hack might be feasible if we use multi-tree-attacks. For , running 2^{15} multi-tree attacks with *m* = 10^{4} takes around 2 minutes and produces an output of 5.2·10^{5} characters. This is still slightly to large for most problems, but could be split up into multiple hacks in an open hacking phase, for example.

In C++11 you can use `mt19937`

and `uniform_int_distribution`

instead of `rand()`

.

### Low-precision time (`Time(NULL)`

)

`Time(NULL)`

only changes once per second. This can be exploited as follows

- Pick a timespan Δ
*T*. - Find an upper bound
*T*for the time you'll need to generate your tests. - Figure out the current value
*T*_{0}of`Time(NULL)`

via custom invocation. - For
*t*= 0, ..., (Δ*T*) - 1, replace`Time(NULL)`

with*T*_{0}+*T*+*t*and generate an anti-test for this fixed seed. - Submit the hack at time
*T*_{0}+*T*.

If your hack gets executed within the next Δ *T* seconds, `Time(NULL)`

will be a value for which you generated an anti-test, so the solution will fail.

### Random device on MinGW (`std::random_device`

)

Note that on codeforces specifically, `std::random_device`

is deterministic and will produce the same sequence of numbers. Solutions using it can be hacked just like fixed seed solutions.

# Notes

- If I made a mistake or a typo in a calculation, or something is unclear, please comment.
- If you have your own hash randomization scheme, way of seeding the rng or anti-hash algorithm that you want to discuss, feel free to comment on it below.
- I was inspired to write this blog after the open hacking phase of round #494 (problem F). During (and after) the hacking phase I figured out how to hack many solution that I didn't know how to hack beforehand. I (un-)fortunately had to go to bed a few hours in (my timezone is UTC + 2), so quite a few hackable solutions passed.

Auto comment: topic has been updated by dacin21 (previous revision, new revision, compare).Suppose that we use some fixed prime

p(for example,p= 10^{9}+ 7, the reason to consider this number is thatp- 1 = 2q, whereq= 5·10^{8}+ 3 is prime, has only one small prime factor: 2) and randomize the base. This clearly rules out your high collision probability example because it only can happen whenn= 3 orn≥q+ 1, neither of this situation is close to the model case whenn≈ 10^{5}. Can we find a bad case in this situation?I have a proof that there are two strings that have same hash with probability at least , where is some constant. However, I don't know any way to construct such two strings in reasonable time. Does anybody know a way (possibly completely different)?

Proof, some math hereConsider — the field of remainders modulo

pand the ring of polynomials over it. Clearly, finding a polynomial in of degree less thannwith coefficients in range [ - (|Σ| — 1), (|Σ| — 1)] with a lot (respectively ton, of course) of roots (in ) is sufficient for finding a hash collision. Letk: = |Σ| - 1) and consider a polynomial (we can take anyrdistinct elements of as roots ofq, not necessarily 0, ...,r- 1), whereris some integer that we will choose later.Consider all polynomials from with degree less than

nand coefficients in range [0,k]. There arek^{n}of them. Ifk^{n}>p^{r}, then two of them have the same remainder moduloq, because possible remainders moduloqare polynomials from our ring with degree <r; their difference is the polynomialPwe were looking for — it is divisible byq, therefore it hasrroots, it has degree less thannand its coefficients are from range [ -k,k].If we choose , then

k^{n}>p^{r}holds, therefore we found a polynomial of degreenwith at leastcnroots, where (probably up to ± 1 to some direction). In the casek= 26 - 1 = 25,p≈ 10^{9}we havec≈ 0.16, which is large enough I guess.Moreover, one may heuristically apply birthday paradox argument here, making

ceven bigger. This does not solve the problem of actually finding polynomials with same remainders moduloq, though.P. S. It is also possible to break single hash modulo 64-bit (link to the problem, it is in Russian though, Breaking hashing). I hope Gassa wouldn't mind explaining the solution, he will definitely do this better than me. In fact, I remember somebody claiming that they can break fixed hashes modulo number of order 10

^{100}./blog/entry/17507?locale=ru#comment-223614

Big thanks to both of you for linking me this, I'll add a subsection on this. (The attack can be improved by storing multiple values at each node; if we store 10

^{4}values per node, the resulting length drops to 64 forp≈ 2^{31}and 256 forp≈ 2^{61}and with 10^{5}values it drops to 32 and 128 and runs in < 1 s.)Thank you very much for your post!

By the way, today I came to the conclusion, that we can calculate double rolling hash without operations

`%`

, if we use two modules: prime`m1 = 2^31-1`

and not prime`m2 = 2^64`

. For hash by`m2`

we just need to work in`unsigned long long`

.Let's find out how to take the remainder of the division by

`m1`

.A simple analogy with the 10-base number systemIf we have number in 10-base system:

`number = digit[0] + digit[1] * 10 + digit[2] * 100 + digit[3] * 1000 + ...`

And want calculate remainder from division by mod

`9`

, we just need to accumulate all digits and repeat while`sum >= 9`

, because`10^p % 9 == 1`

`number % 9 = (digit[0] + digit[1] + digit[2] + ...) % 9`

This works for

`9 * k`

not as you expect. This method gives the result from`1`

to`9`

for all positive values and`0`

only for`0`

Let

`x`

be from`1`

to`(m1-1)^2`

— result from multiplication of two remainders. For getting remainder modulo`2^31-1`

we need to set`x`

a sum all digits in base`2^31`

and repeat while`x >= 2^31-1`

.But it works for

`(2^31-1) * k`

numbers not as you expect. This method gives the result from`1`

to`2^31-1`

for all positive values and`0`

only for`0`

.`x % (2^31-1)`

is number from`[0..2^31-2]`

This method definitely gives number from

`[1..2^31-1]`

.It works great for rolling hashes if we cut

`0`

and if we compare results from one type of method (works only with remainders`[1..2^31-1]`

against`[0..2^31-2]`

). Maybe it can speed up rolling hashes solutions.Thanks a lot for your comment.

First of all,

don't use2^{64}as your modulo! It can be hacked regardless of the base. (See the section 'Hashing with unsigned overflow (p= 2^{64}, q arbitrary)'). Combining it with another hash does not save you, see the section 'Multiple hashes', the Thue–Morse sequence can be used there as well.Second of, your bit-trick technique can be used to map to as well, just add 1 to

xbeforehand and subtract 1 fromxafterwards. It is probably more useful if we usep= 2^{61}- 1 (with some more bit-tricks), as the alternative with`__int128`

does not exists on codeforces.Badly tested code for that pBasic idea: Split the number into lower 32 and upper 29 bits. Then multiply the parts

^{64}- 2^{33}+ 1, so we do one reduction on it.^{29}- 1)^{2}·8 < 2^{61}.)I'll add this (with

p= 2^{61}- 1) to the section about larger modulos once I get around to test and benchmark it.Thanks, please tell me how fast hacker can hack my solution if he knows:

`p1 = 2^31-1`

and`p2 = 2^64`

`(256, 2^31-1)`

with`std::random_device`

,`std::mt19937`

,`std::uniform_int_distribution`

`2^64`

exist anti-hash test regardless of the base`q`

`14:52:31`

.At first I misread

`std::random_device`

as`high_resolution_clock`

and it seemed quite hard. (The main issue is to breakp_{1}with less than characters.) But on codeforces, the former always returns`3499211612`

as the first number. I can hack you in less than a minute as followsp_{1}(takes < 1 second).You need to combine

`sqrt(p1) ~=~ 2^16`

chars with`2^11`

of Thue–Morse string, this is`2^27 = 134.217.728`

chars, if I understand correctly, it is possible? I never seen input greater than`5 * 10^6`

chars.I asked it because I wrote tutorial for beginners last day and in this tutorial I used two modules:

`2^64`

,`10^9+123`

and random point, generated with`std::mt19937`

,`std::chrono::high_resolution_clock`

and`std::uniform_int_distribution`

and wrote that using`std::time(0)`

or`std::random_device`

in MinGW not safety, but now I understand that you need very big input to hack solution with fixed point and maybe don't need randomization?You're mixing apples and oranges. For the birthday attack, we need to generate

strings. These strings can be quite short, in practice around 6characterseach. We then picktwostrings that hash to same value and use them in the Thue-Morse sequence. So the total length would be 6·2·2^{10}= 12288. (2^{10}sufficies, 2^{11}was based of a non-tight bound from the linked post on hashes mod 2^{64}.)If you use 2

^{64}and a prime around 10^{9}with good randomization, I haven't figured out a good hack yet. (Note that this does not mean you can't be hacked, someone might find a better attack.) The best I thought of was building a generalized Thue-Morse sequence in base , so that all pairs collide mod 2^{64}and there is a good change for a birthday-collision modp. This would lead to massive input however (Estimated 2^{10}·6·45'000 ≈ 3·10^{8}), as it involves submitting an input with long strings. So I guess there's a good chance you won't be hacked, but I myself prefer something where I can proof that it's hard for me to get hacked and not rely on other not figuring out some new technique.Now I understood, thanks

I not understood why this method not works on this problem on generated test. Code on ideone.com or my solution is bad.

I generated test with length 5000:

True answer is 2500, but it gets 2499 or 2500 (Answer depends on launch with random point).

UPD: Sorry, bug in function diff, my bad:a and b unsigned long long and it is not correct (compare with zero) in function sub:

`return (a -= b) < 0 ? a + mod : a;`

It should be like this:

`return (a -= b) >= mod ? a + mod : a;`

My randomization function:

Thanks, added it and a version that uses the corresponding built-in function instead of asm.

Double hashing has never failed me. So I think you don't have to worry for collision in CE if you use double hashing. My primes for mod are 10

^{9}+ 7 and 10^{9}+ 9 and I use two random primes like 737 or 3079 or 4001 as base.If the input is random so that we can assume the hash-values are uniformly distributed, you're fine.

But if you don't randomize your bases at runtime (and it sounds like you don't), someone might hack your solution. (And AFAIK it makes no difference whether the

baseis prime or not.)e-maxx has it mentioned that we should choose base as prime > max character in string. It says for eg. if string is all lowercase choose 31. But it didn't mention why.

Now that you are no more random :D, Get ready to... :D

Auto comment: topic has been updated by dacin21 (previous revision, new revision, compare).On abusing multiple known modulos: you can

quicklygenerateshorthack-tests using lattice reduction methods (LLL). There was a related challenge at TokyoWesterns CTF (an infosec competition) and I made a writeup for that (or another writeup). There you have 8 32-bit modulos and the generated hack-test is less than 100 symbols. (The problem there is palindrome search but the method is generic).Thank, lattice reduction works great for multiple hashes if the alphabet size is not to small. I added a section on it with some modified code. For many 61-bit hashes, the LLL algorithm deteriorates, but the BKZ algorithm in sage works instead.

Thanks for your awesome post !

Just another suggestion or request: it would be great if you would provide a sample full code combining all of the ideas at the end of the post.

Awesome post. .simple and really informative..!!

Although it sounds rather weird to bump such an old blog, I want to point out that you can actually perform a birthday attack with $$$O(1)$$$ memory using Floyd's cycle finding algorithm (similar to pollard-rho).

Assuming your hash function is $$$H: M\rightarrow D$$$, let's find a random mapping $$$F: D \rightarrow M$$$ that generates a random message using a hash value as seed. Assume $$$F$$$ is injective.

Let $$$x:=1, y:=1$$$ in the beginning, $$$T(x)=H(F(x))$$$ and repeat $$$x:=T(T(x)), y:=T(y)$$$ until $$$x=y$$$. We've found a collision and we need to retrace how it happened. This can be done by resetting $$$x:=1$$$ and repeat $$$x:=T(x), y:=T(y)$$$ until $$$T(x)=T(y)$$$. $$$(F(x),F(y))$$$ will be such a collision pair.

CorrectnessAssume you've understood the correctness of pollard-rho. Notice that by the end of the first phrase, $$$x$$$ is faster than $$$y$$$ by some multiple of cycle length, thus the round passed for $$$y$$$ is also some multiple of cycle length. Therefore, we can keep that advantage and trace from the beginning back again.

Sample ImplementationThere's also a quicker (around 2x) way that uses $$$O(|D|^{1/4})$$$ space. Still assume $$$T(x)=H(F(x))$$$. We still iterate through $$$x,T(x),T(T(x)),T(T(T(x))) \cdots$$$, but only record some interesting entries (e.g. entries that are multiples of some integer $$$G\geq |D|^{1/4}$$$) in the hash table. When a collision in the hash table is found, we trace back that 'gap' and find out the real collision. You may refer to the code for more details.

Sample ImplementationWhen using multiple hashes, is it necessary to use different modulo, or is it ok to just a different base?

Update: It's ok just to change the base.