### cfalas's blog

By cfalas, 16 months ago,

Hello CF! Following IOI 2022, I did some analysis from the scoreboard. Here are my findings.

## Country Standings

This has already been posted on codeforces here, but mine is slightly different as it uses the total score instead of the average (hence countries with less than 4 participants are ranked higher than the ones with 4 in mine).

Top 5:

Rank Country Average Score
1 CHN 577.50
2 USA 477.14
3 JPN 467.77
4 KOR 406.11
5 CAN 376.01
Full Country Standings

## Problem Score Distributions

Most people do better on some tasks than on others. These graphs can answer the following questions:

• What medal would I get if the score was entirely based on one task?
• Which was the easiest task to get x pts on?
• How many people got >x pts on each task? etc (You can see a higher resolution version here)

The vertical coloured lines represent the "cutoff" score for each medal for each task (ie bronze line corresponds to median score, silver corresponds to 75th percentile, and gold corresponds to 91.67th percentile.

## Country Score Dispersion

Following day 2, some people found that some online contestants from the same country had similar scores a bit "sus". In this graph, we can see the distribution of the coefficients of variation of the scores of each country's contestants. ie for each country:

• Take the score of the 4 contestants.
• Find the standard deviation and mean
• Find the coefficient of variation (standard deviation / mean)

Plot the distribution of said coefficients over all countries, separately for online and onsite ones.

Note: the distribution has been estimated using a guassian KDE, which is why there are probabilities >0 for values of x < 0. The actual smallest coefficients are 0.0287 for onsite participants (JPN) and 0.0450 for online participants (CHN).

Note that despite the fact that the average variation is less for online participants, the sample size for them is quite small (n=17).

## Sources

These were generated using data from the scoreboard, available at https://ranking.ioi2022.id.

Online/Onsite participants were found at https://status.ioi2022.id

• +74

By cfalas, history, 4 years ago,

In response to blogs such as this and this, I have made a similar graph for IOI 2019. I was interested in finding how I could expect to do if I qualified, and the only metric I have at the time is CF rating (since there isn't anywhere to submit the problems as of now)

Spoiler
Spoiler 2

So, here it is:

Here are some things to take into consideration:

• A lot of the participants didn't have their CF account on stats.ioinformatics.org. These participants were dismissed when making the graph
• The ratings of the participants were taken at present time

As you can see, as in previous IOI demonstrated by the blogs above, even though there is a clear trend, there's too much variation in the data to get something out of. However, here are some interesting facts:

• Out of the people that got gold, the one with the lowest rating, has a lower rating than the person that has the lowest rating out of the silver medals.
• Only 30.58% of the participants have a linked CF account(!)
• Mean ratings for each medal: (only with the ratings of the participants that had their account listed on stats.ioinformatics.org)
• Gold: 2465.5625
• Silver: 2248.24
• Bronze: 2026.125
• No Medal: 1799.2
• Does this mean that people with lower ratings can't get a medal?
• They can, but it's not as easy. The lowest ratings for each medal are:
• Gold: 1883
• Silver: 1932
• Bronze: 1597
• No Medal: 1223

Sources:

If there are any issues with the statistics please let me know

• +118