From: El Kot on
milivella wrote:
> El Kot:
>
>> We can use the interim EPL table to judge how good FM's simulation
>> is. Frankly, I'm not impressed. You can compare the simulated table with
>> the real one at Kicker:
>>
>> http://www.kicker.de/news/fussball/intligen/england/barclays-premier-...
>>
>> One would expect the Fulham results to be different from Fulmark's, but
>> the rest of the teams should have performed in simulation more or less
>> similar to reality.
>
> You are highlighting an interesting point. Leaving aside some banal
> considerations (e.g. a. the comparison will be most interesting at the
> end of the season, after 380 matches; b. you can have a bad
> simulation, or no simulation at all, and still produce a table that is
> similar to reality), I have some questions:

Hehe, your point b) is funny. One judges a simulation by its
outcome; if a simulation produces a result that is the same as what it
simulates, then it is, by definition, good. :)
Here is what I think about the points you raise:

- the real match outcomes and table are a sampling of the teams' real
strengths. They are, of course, subject to randomness; individual
matches can be influenced by chance. However, repeated sampling will
converge to the true value, because I don't think the influence of
chance is /that/ big that it dominates; the variance is not that big.
Remember that a sequence is a product of chances; you can't expect the
unexpected to happen too often in a row.

- the simulation has the advantage that it can be repeated and its
statistics checked out. If and when you have time, please make several
runs, and check the following things:
1. How similar to each other the simulated tables are at the end of
the season. This will tell us how consistently the FM engine applies the
parameters of the players and teams. If the tables turn out to be more
or less similar (with each team moving by no more than, let's say, 2 to
3 positions around the table), we can conclude that it is a consistent
engine.
2. Then we can check how true to life it is, by comparing the actual
position from real life to the simulated ones. This will tell us how
realistic the parameters and methods it uses are.
3. And of course we can see how it is influenced by randomness, by
looking at the outcomes of one match played several times. This is
allowed to have a bigger variance, but still within limits. One wouldn't
expect ManU to beat Portsmouth less than 8 times out of 10.

- what you say below, is backwards. We can't really calculate the
statistics of the real teams; there's too few samples. We can, however,
calculate the statistics of the simulation, and then see if the real
results fall within the predicted confidence intervals. And that will
tell us how true to life the simulation really is - if the predicted
points per game for ManU is x points, and the actual one is y, we can
check to see within which confidence interval of the simulation that
value falls, and deduce the quality of the simulation. I'm not a
statistician, so I will abstain from claiming that, if the actual
results of the teams fall within the 80% confidence intervals, it means
that the simulation's accuracy is 80%. The relationship is surely more
complicated; if Daniele or someone else (MH?) knows more, please share.


> 1. Is it correct to consider the real EPL table like a sample of
> teams' real value (rather than a perfect representation of it)? After
> all, one could say that, if Blackburn and Chelsea play 10 times,
> Chelsea would win 6 matches (random example), so their draw is like 1
> of the other 4 matches, i.e. it's a sample of the 10 hypothetical
> matches between the two teams.
>
> 2. If the answer to (1) is "yes": we can compute a confidence
> interval, can't we? I.e. we can say "with a confidence level of 95%,
> Manchester United's real value measured in points per game is 2.23 (=
> 69 points / 31 matches) plus or minus x".
>
> 3. Things are of course easier in FM, because there we can simulate
> the same season a lot of times (I just have to remember to check the
> game after a couple of hours, save the standings and re-start the
> season). After having ran a lot of simulations and having averaged the
> results, we should be able to check if the simulated results are
> inside or outside the confidence interval mentioned at (2), so
> verifying/falsifying (in the limits of what it possible) FM's claims
> of realism. Am I wrong?

--
No, no, you can't e-mail me with the nono.
From: milivella on
milivella:

> 1st update: September 16: Fulham is 18th.

2nd update: October 16: Fulham is still 18th.

Data-filled screenshots:
Played matches: http://i.imgur.com/FN4Y7.jpg
Competitions:
- Premier League: http://i.imgur.com/v6F3j.jpg
- Europa League: http://i.imgur.com/lNCbJ.jpg
- [out of League Cup]
Players' stats:
- Goalkeepers: http://i.imgur.com/v8YV9.jpg
- Defenders: http://i.imgur.com/8jvvU.jpg
- Midfielders: http://i.imgur.com/3xgwt.jpg
- Attackers: http://i.imgur.com/u70rm.jpg

--
Cheers
milivella
From: Jesus Petry on
On Mar 25, 11:09 am, milivella <milive...(a)gmail.com> wrote:
> milivella:
>
> > 1st update: September 16: Fulham is 18th.
>
> 2nd update: October 16: Fulham is still 18th.
>
> Data-filled screenshots:
> Played matches:http://i.imgur.com/FN4Y7.jpg
> Competitions:
> - Premier League:http://i.imgur.com/v6F3j.jpg
> - Europa League:http://i.imgur.com/lNCbJ.jpg

But I see they have games in hand and are doing well in Europe!

Tchau!
Jesus Petry
From: milivella on
El Kot:

Thanks for the reply.

>      Hehe, your point b) is funny. One judges a simulation by its
> outcome; if a simulation produces a result that is the same as what it
> simulates, then it is, by definition, good. :)

OK. Question: if, at the beginning of the season, you ask me
tiopredict the final table and I just copy the previous season's final
table (replacing relegated teams), is it a simulation? (more questions
to come if you answer "yes"...)

> - the real match outcomes and table are a sampling of the teams' real
> strengths. They are, of course, subject to randomness; individual
> matches can be influenced by chance. However, repeated sampling will
> converge to the true value, because I don't think the influence of
> chance is /that/ big that it dominates; the variance is not that big.
> Remember that a sequence is a product of chances; you can't expect the
> unexpected to happen too often in a row.

Of course. I specified it to highlight two things:
- You can't judge a simulation using just one or few real matches as
benchmark.
- If a simulation has Chelsea at 70 points and Chelsea really has 69
points you can't just say that the simulation is wrong. (this was
functional to my idea to deduce teams' strength from the real table,
idea that thank to you now I know is wrong)

> - the simulation has the advantage that it can be repeated and its
> statistics checked out. If and when you have time

I just have to develop the habit to start a new simulation before I go
to sleep, and the morning after I'll find the final table (indeed I
guess that it doesn't take more than a couple of hours to complete a
season, probably less). Write down the number of points for each team,
and it's done. (tl;dr: it takes no time)

> please make several
> runs

How much is "several"?

> and check the following things:
>    1. How similar to each other the simulated tables are at the end of
> the season. This will tell us how consistently the FM engine applies the
> parameters of the players and teams. If the tables turn out to be more
> or less similar (with each team moving by no more than, let's say, 2 to
> 3 positions around the table), we can conclude that it is a consistent
> engine.

What's the metric? Variance of number of points (or: of position) for
each team?

>    3. And of course we can see how it is influenced by randomness, by
> looking at the outcomes of one match played several times. This is
> allowed to have a bigger variance, but still within limits. One wouldn't
> expect ManU to beat Portsmouth less than 8 times out of 10.

Hmm... If it's just *one* match each season (i.e. if every morning
I'll have to check just ManU vs. Portsmouth), it's OK. Otherwise it's
not practical, because there isn't (AFAIK) a way to export results
from FM to a spreadsheet.

> - what you say below, is backwards.

Thanks for the explanation!

> We can, however,
> calculate the statistics of the simulation, and then see if the real
> results fall within the predicted confidence intervals. And that will
> tell us how true to life the simulation really is - if the predicted
> points per game for ManU is x points, and the actual one is y, we can
> check to see within which confidence interval of the simulation that
> value falls, and deduce the quality of the simulation.

I.e. after several runs of the simulation we can say that we know each
team' strength (as measured by points per match) according to FM
(parameter), and we check how many standard errors separate the real
EPL table (sample) from it. I.e. we ask: how likely is that the real
table is generated by FM?

Am I right?

Sorry for questionable lexicon: all my statistical knowledge comes
from my current reading of Statistics for Dummies*... and I'm
understanding just half of what I am reading!

* Really! I can't keep annoying Daniele for every little doubt, after
all.

--
Cheers
milivella
From: milivella on
milivella:

> 2nd update: October 16: Fulham is still 18th.

3rd update: November 16: Fulham is 15th.

Data-filled screenshots:
Played matches: http://i.imgur.com/dj1ai.jpg
Competitions:
- Premier League: http://i.imgur.com/5qfzS.jpg
- Europa League: http://i.imgur.com/tDHBS.jpg
- [out of League Cup]
Players' stats:
- Goalkeepers: http://i.imgur.com/XGGz2.jpg
- Defenders: http://i.imgur.com/9J7br.jpg
- Midfielders: http://i.imgur.com/f5UZY.jpg [this URL is an evident
homage to the Mothers of Invention]
- Attackers: http://i.imgur.com/ZeHDb.jpg

--
Cheers
milivella
First  |  Prev  |  Next  |  Last
Pages: 1 2 3 4 5 6
Prev: Volkswagen - Golden Horde [R]
Next: Brasileirão Maluco 2010