Excelsior

Follow us on

excelsiorathletic@gmail.com

07976 306 494

Main Menu

Latest Blog Entry

Reflections from a Gymnastics assistant coaching course
1st June 2023
A guest post from Kath Maguire. Kath is the parent of one of our club’s gymnasts. She asked about doing some volunteering a couple of months ago and whether there was a course she could do. Here are her reflections from the day. “I’ve been thinking about volunteering for a while now but as it’s […]
More

User login

Lost password?

Improper Application and Interpretation of Sports Science Statistics

Juking the Stats: Why not all “research” is valid

The latest craze in competitive sport appears to be the use of data to aid understanding of, and improvement in sporting performance. This has resulted in a glut of material, each item claiming to have established some new result which may have useful implications in the development and performance of human athletes.

There are often studies conducted with non-athletes as well, and the line between what could be considered medical research as opposed to what is known as sports science is not clearly defined.

I should stress that my knowledge of the field of sports science is limited, the purpose of this article is to question the structure and findings of some typical articles.

A typical paper in this field might take the following form:

1) Design a study with some hypothesis of interest
2) Collect data from subjects (fitness testing)
3) Analyse the data to check for consistency with the hypothesis
4) Draw conclusions.

 A good statistician should be able to perform multiple roles.

In my opinion, some of the most important are:

  1. To decide on the real questions at the heart of a problem of interest, not to just churn out results for the sake of it.
  2.  To decide if a hypothesis is necessary, and if so to construct one which is of real actual interest. Sometimes it is best to approach a problem with an open mind, in the knowledge that there are likely to be interesting results, but unsure of what they will be.
  3. To employ appropriate methods (typically statistical models) to analyse the collected data (we don’t need to get too technical here).
  4. To explain the underlying reasons behind any results – studies in which results are simply quoted as gospel are of limited interest to me.
  5. To critically review the work, pointing out potential shortcomings and areas for future research.

The final point is perhaps the most interesting. It is often the role of a statistician to dampen (or in some cases pour cold water) on enthusiasm about some exciting results.

Sports Science Statistics must be taken in context.

Conclusions drawn from a study of, say, a weightlifter’s improved performance due to a certain type of training programme should not be used as an automatic basis for a different strength-based sport, such as rowing.

I work in the field of weather forecasting. A modern-day weather forecast involves running a computer model forward in time to produce a single forecast of the atmosphere. Statistics of this forecast (such as the average forecast error) can be calculated at different locations. It is well-known that such statistics vary by location – it is more difficult to predict the weather in Shetland than in the Sahara Desert. We could not, therefore, use statistics derived from one location to predict the average forecast error in another.

In short, statistics is about describing what might have happened in a given context, but didn’t. We can use these findings to issue probabilities of what might happen in the future, on the basis that the context is consistent.

Forget the weather: what about sport science?

The few articles I have read in the sports science field (in all honesty I couldn’t face reading too many!) seem to fall short on many of the above points.

For example, Owen et al. (J Strength Cond Res, 2011) conduct a study of heart rate responses of soccer players when playing in three-sided and nine-sided games. They conclude that the HR of players in three-sided games is consistently higher than for nine-sided games. They also note that three-sided games provide more shooting chances, and encourage players to run more with the ball, whilst the nine-sided games produced more tackles, passes and interceptions.

They draw the conclusions that three-sided games are preferable for fitness training, and suggest that strikers should participate in three-sided games whilst defenders should concentrate on nine-sided games.

I have two main problems with this work from both a scientific and practical viewpoint.

  1. The statistics quoted in themselves should be treated with caution, given the small sample size of fifteen players who participated in only a few games of each type. Without conducting a formal test I cannot be more precise, but these measurements are undoubtedly subject to substantial variation.
  2. What insight does the study really offer us? Aren’t the findings, on which the entire article is based, merely confirmation of the obvious? It is useful here to consider the so-called `pyramid of outcomes’ .

This study gives only surrogate measures (the base of the pyramid), but assumes in the conclusion that such surrogates automatically extend in to true performance measures (essentially whether they can be used to increase the probability of winning football matches).

This assumption seems completely without foundation when one considers the practical implications of the study. For example, suppose that on the basis of the study, strikers train in three-sided games whilst defenders train in nine-sided games, in order to provide more shooting opportunities for strikers and more defending opportunities for defenders. Is there really any point in this? Wouldn’t three-sided games just result in strikers shooting from anywhere, and playing (by definition) against less able defensive opposition? Surely the way to improve as a striker is to learn how to play against good defenders?

Frankly, this work smacks of conducting a study for the sake of it, and drawing conclusions based on a few surrogate measurements without paying any attention to the sport of interest.

How to conduct a more informative study.

1) Collect a larger sample of players from a variety of clubs, preferably from different countries.
2) Train different groups of players in different environments, as suggested by the study.
3) Collect surrogate measurements from the different training sessions.
4) Examine if the surrogates had an effect on actual game results (i.e. construct a proper statistical model rather than merely reporting surrogate values).
5) Examine whether a return to previous training routines result in a reversion to previous performance.

A statistical model is essentially the use of surrogate measurements to aid in predicting the value of, and assessing the uncertainty in, measurements at the top of the pyramid. The article mentioned here simply assumes that larger surrogate values immediately imply improved results, an assertion which is without foundation.

Such a study would admittedly be hard to carry out both practically and from a theoretical statistical viewpoint. However, we are dealing with complicated situations – we are essentially trying to model outcomes from the human body, an immensely complicated organism.

This is my overriding point, studies which simply churn out results for the sake of publishing papers are of little practical use. I would go further and suggest that they are actually dangerous in the wrong hands – a statistical model is no good in the hands of an incapable operator.

Conclusion

From my brief consultation of the literature, I have seen many examples of a mis-use of statistics which would not be permitted in a statistics journal.

The typical methods used are likely to underestimate the complexity of the situation at hand. I suspect therefore, that the true value of statistics such as the p-value are somewhat larger than reported.

I feel confident in ascerting that the conclusions of the articles I have read are based on extremely shaky ground in a theoretical sense, let alone their practical shortcomings.

Robin Williams Statistics Phd Student (University of Exeter),  England Blind Footballer, 2012 Olympian

More on interpretation of data here 

Comments

  1. […] getting bogged down in the statistics, there was no significant difference between the actual 1-RM and the predicted 1-RM, and the […]

  2. […] evidence and opinion. Evidence can be gained from personal observation and experience, and also the correct interpretation of scientific research, as well as experience of other […]

  3. […] Improper application and interpretation of statistics […]

  4. […] part of my preparation for the workshop I trying to find research that was both current and measured what we are trying to train (rather than some abstract concept […]

  5. […] Improper application and interpretation of statistics in sports “science” […]

  6. Topsy says:

    Well put and informative James, thank you.
    It can be a bit of a minefield these days, with is much information being pushed out there, its nice to see some clarity on it.

Leave a Reply

Your email address will not be published. Required fields are marked *

This site uses Akismet to reduce spam. Learn how your comment data is processed.

Client Testimonials

Josh Steels: wheelchair tennis
I started working with James 3 years ago via the TASS programme. When James first met me, physically I was nowhere the best I could be. Since working with James I have seen vast improvements in my fitness and strength which has been put into great use on court. Each session is worked around making sure I am able to get the best quality training as well as catering for my chronic pain and fatigue levels. On top of this James has always been happy to meet at facilities that are best for myself meaning I could fit training sessions in on route to tournaments or camps.
 
More

Upcoming Courses