Article Text

Download PDFPDF

Reductions in pre-season training loads reduce training injury rates in rugby league players
  1. T J Gabbett
  1. Queensland Academy of Sport, QLD, Australia
  1. Correspondence to:
 Dr Tim Gabbett
 Queensland Academy of Sport, PO Box 956, Sunnybank, QLD 4109, Australia; tim.gabbettqld.gov.au

Abstract

Objectives: To investigate if reductions in pre-season training loads reduced the incidence of training injuries in rugby league players, and to determine if the reductions in training loads compromised the improvements in physical fitness obtained during the pre-season preparation period.

Methods: A total of 220 sub-elite rugby league players participated in this 3 year prospective study. Players underwent measurements of speed, muscular power, and maximal aerobic power before and after three 4 month (December to March) pre-season preparation periods (2001–2003). A periodised skills and conditioning program was implemented, with training loads progressively increased in the general preparatory phase of the season (December to February) and reduced slightly in March in preparation for the competitive phase of the season. Training loads were calculated by multiplying the training session intensity by the duration of the training session. Following the initial season (2001), training loads were reduced through reductions in training duration (2002) and training intensity (2003). The incidence of injury was prospectively recorded over the three pre-season periods.

Results: The training loads for the 2002 and 2003 pre-season periods were significantly lower (p<0.001) than those in 2001. The incidence of injury was significantly higher in the 2001 pre-season than the 2002 and 2003 pre-season periods. The increases in maximal aerobic power progressively improved across the three seasons with a 62–88% probability that the 2002 and 2003 pre-season improvements in maximal aerobic power were of greater physiological significance than the 2001 pre-season improvements in maximal aerobic power.

Conclusions: These findings demonstrate that reductions in pre-season training loads reduce training injury rates in rugby league players and result in greater improvements in maximal aerobic power.

  • MCID, minimum clinically important difference
  • RPE, rating of perceived exertion
  • conditioning
  • injury prevention
  • performance
  • rugby league
  • sub-elite

Statistics from Altmetric.com

Request Permissions

If you wish to reuse any or all of this article please use the link below which will take you to the Copyright Clearance Center’s RightsLink service. You will be able to get a quick price and instant permission to reuse the content in many different ways.

Rugby league is an international collision sport played at amateur,1,2 semi-professional,3,4 and professional levels.5–8 The game is physically demanding requiring players to compete in a challenging contest involving frequent bouts of high intensity activity such as running, passing, and sprinting, separated by short bouts of low intensity activity such as walking and jogging.9 During a match, players are exposed to numerous physical collisions and tackles.8,10 As a result, musculoskeletal injuries are common.11,12

While several studies have documented the incidence of rugby league playing injuries,5–7,10–16 training injuries in rugby league have received relatively little attention.17,18 Recent evidence has shown that the majority of rugby league training injuries occur in the pre-season preparation period (December to March) when training loads are greatest.18 Indeed, the pre-season incidence of injury of 116.1 per 1000 training hours is 2.6-fold higher than the seasonal average injury rate (45.3 per 1000).14 Moreover, increases in training loads are significantly correlated (r = 0.86) with increases in training injury rates.18 These findings suggest that the harder rugby league players train, the more injuries they will sustain and that reductions in training loads will reduce training injury rates in these athletes. However, due to the physical nature of the sport, rugby league training sessions require an adequate intensity in order to develop the required physical qualities needed to compete optimally for the duration of a match. Therefore, reductions in training loads could minimise the pre-season improvements in physical fitness that are required for the forthcoming playing season.

Effective prevention of sporting injuries is dependent on the identification of the extent of the injury problem and the causes of injury, the implementation of injury prevention strategies, and the re-evaluation of the extent of the injury problem.19 While several studies have documented the extent of the injury problem in rugby league, an evidence based injury prevention program has not been implemented for this sport. With this in mind, the purpose of the present study was to investigate if reductions in pre-season training loads reduced the incidence of training injuries in rugby league players. In addition, a secondary purpose of this study was to determine if the reductions in training loads compromised the improvements in physical fitness obtained during the pre-season preparation period.

METHODS

A total of 220 healthy men, registered with the same sub-elite rugby league club participated in this 3 year prospective study (2001–2003). The total number of registered players over the 3 year period was 79, 65, and 76, respectively. Eleven (5.0%) players played two or more seasons. All players were competing within the Gold Coast Group 18 (New South Wales Country Rugby League, Australia) (2001–2002) or South-East Queensland (Queensland Rugby League, Australia) (2003) senior rugby league competitions. Each pre-season training period lasted from December through March, with the competitive season lasting from March through September. All players underwent fitness testing in December and March as part of their pre-season training program.

Fitness testing battery

Muscular power (vertical jump),20 speed (10 m, 20 m, and 40 m sprint),21 and maximal aerobic power (multi-stage fitness test)22 were the fitness tests selected. The age and playing experience of players was also documented. Players were instructed to refrain from strenuous exercise for at least 48 h prior to the fitness testing session and consume their normal pre-training diet prior to each testing session. At the beginning of the fitness testing session, players underwent a standardised warm up (progressing from low to higher intensity activity) and stretching routine. Players performed two trials for the speed and muscular power tests, with a recovery of approximately 3 min between trials. Players were encouraged to perform low intensity activities and stretches between trials. Upon completion of the speed and muscular power tests, the field testing session was concluded with players performing the multi-stage fitness test (estimated maximal aerobic power).

Muscular power

Lower leg muscular power was evaluated by means of the vertical jump test.20 Vertical jump height was calculated as the distance between the highest point reached during standing and the highest point reached during the vertical jump. Vertical jump height was measured to the nearest 1 cm with the highest value obtained from two trials used as the vertical jump score. The intraclass correlation coefficient for test-retest reliability and typical error of measurement for the vertical jump test was 0.96 and 3.3%, respectively.

Speed

The running speed of players was evaluated with a 10 m, 20 m, and 40 m sprint effort21 using dual beam electronic timing gates (Speed Light Model TB4, Serial No. 4921001; Southern Cross University Technical Services, Lismore, Australia). Speed was measured to the nearest 0.01 s with the fastest value obtained from two trials used as the speed score. The intraclass correlation coefficient for test-retest reliability and typical error of measurement for the 10 m, 20 m, and 40 m sprint tests were 0.95, 0.97, and 0.97, and 1.8%, 1.3%, and 1.2%, respectively.

Maximal aerobic power

The multi-stage fitness test was used to estimate maximal aerobic power (V˙o2max).22,23 A correlation of 0.92 has been reported between the level achieved during the multi-stage fitness test and treadmill determined V˙o2max.23 In addition, all players completed duplicate multi-stage fitness tests, performed 1 week apart, prior to the commencement of this study to determine test-retest reliability. The intraclass correlation coefficient for test-retest reliability and typical error of measurement for the multi-stage fitness test were 0.90 and 3.1%, respectively.

Training loads

Each player participated in two organised field training sessions per week. A periodised, game specific training program was implemented, with training loads progressively increased in the general preparatory phase of the season (December to February), and reduced slightly in March in preparation for the competitive phase of the season. The duration of training sessions was recorded, with sessions typically lasting between 60 and 100 min. Players participated in 30 pre-season training sessions each year. Based on the finding that training injury rates are increased with increases in training loads,18 following the initial season (2001), training loads were reduced through reductions in training duration (2002) and training intensity (2003).

The intensity of individual training sessions was estimated using a modified rating of perceived exertion (RPE) scale.24 Training loads were calculated by multiplying the training session intensity by the duration of the training session. Intensity estimates were obtained within 30 min of completing the training session. In the present study, the correlation between training heart rate and training RPE, and training blood lactate concentration and training RPE was 0.89 and 0.86, respectively. A subset of players (n = 11) also completed two identical off-season training sessions, performed 1 week apart, prior to the commencement of the study to determine the test-retest reliability of the RPE scale. The intraclass correlation coefficient for test-retest reliability and typical error of measurement for the RPE scale were 0.99 and 4.0%, respectively.

Environmental conditions

Daily weather variables were prospectively measured by the Bureau of Meteorology. Environmental data were taken from the nearest meteorology station (Southport, Queensland, Australia) to the location studied. Maximum, minimum, and average temperature, relative humidity, and rainfall were the environmental variables recorded. Given that a low annual rainfall has been shown to be a significant predictor of injury,25 the 365 day rainfall prior to the commencement of the pre-season training periods was also recorded.

Definition of injury

A head trainer, employed by the club to provide injury prevention and management services, and skills and conditioning coaching, assessed all injuries. The head trainer held tertiary qualifications in exercise and sport science and was nationally accredited in injury prevention, assessment, and management. For the purpose of this study, an injury was defined as any pain or disability suffered by a player during a training session, and subsequently assessed by the head trainer during, or immediately following, the training session.17 All injuries sustained during training sessions were recorded. The severity of injury was classified as transient (no training missed), minor (one training week missed), moderate (two to four training weeks missed), or major (five or more training weeks missed).11,17

Classification of injury

Injuries were categorised according to the site of injury. The head and neck, face, abdomen and thorax, shoulder, arm and hand, thigh and calf, knee, ankle and foot, and “others” categories were the sites selected. Injuries were also described according to the type (nature) of injury sustained. Muscular injuries were classified as either haematomas or muscular strains. Additional categories for the type of injury included joint sprains, concussion, contusions, abrasions, blisters, lacerations, fractures and dislocations, and “others”. Finally, injuries were described according to the cause of injury. Causes of injury were categorised as being tackled, while tackling, being struck by another player or ball, collision with another player or fixed object (for example, goal post, ground), fall/stumble (for example, rolling ankle while running), slip/trip (for example, slipping on wet surface), twisting to pass or accelerate, scrum contact, overexertion (for example, due to rapid changes in speed, intensity, and/or direction), overuse (for example, due to repetitive loading), temperature related disorders (for example, heat stress), and “others”.

Statistical analysis

Differences in the age and playing experience of players, and environmental conditions were determined using a one way analysis of variance. Changes in muscular power, speed, V˙o2max, and training loads over the three pre-season preparation periods were compared using a two way (season×month) analysis of variance and by comparing the true change in performance with the minimum clinically important difference (MCID) for that variable.26 The MCID was defined as the smallest worthwhile change perceived to be physiologically significant to the average athlete. The MCID for muscular power, 40 m speed, and V˙o2max was calculated as 1.5 cm, 0.10 s, and 0.8 ml kg−1 min−1, respectively. Injury exposure was calculated by multiplying the number of players by the session duration. Injury rates were calculated by dividing the total number of injuries by the overall training injury exposure. Expected injury rates were calculated as described by Hodgson Phillips et al.12 The chi-squared (χ2) test was used to determine whether the observed injury frequency was significantly different from the expected injury frequency. All data were reported as means and 95% confidence intervals (CI) and the level of significance was set at p<0.05.

RESULTS

Age and playing experience

The age and playing experience of the players are shown in table 1. Players in the 2001 season had significantly greater (p<0.05) playing experience than those in the 2002 and 2003 seasons.

Table 1

 Age and playing experience of sub-elite rugby league players

Environmental conditions

There were no significant differences (p>0.05) among pre-season periods for the recorded maximum, minimum, and average temperature, and relative humidity. The average rainfall for the 2002 pre-season period (73.9 mm, 95% CI: 35.7 to 112.0) was lower than the average rainfall for the 2001 (184.5 mm, 95% CI: 99.0 to 269.9) and 2003 (181.3 mm, 95% CI: 0.0 to 378.5) pre-season periods. The 365 day rainfall prior to the commencement of the pre-season training period was not significantly different (p>0.05) among seasons.

Training loads

There were significant differences among seasons for training intensity, training duration, and training loads. The training intensity in the 2003 pre-season was significantly lower (p<0.01) than the 2001 pre-season. There were no significant differences between the 2001 pre-season (4.20 units, 95% CI: 4.09 to 4.32) and the 2002 pre-season (4.05 units, 95% CI: 3.89 to 4.22) or the 2002 pre-season and the 2003 pre-season (3.90 units, 95% CI: 3.75 to 4.05) for training intensity. Training duration was significantly higher (p<0.001) in the 2001 pre-season (78.1 min, 95% CI: 77.1 to 79.1) than the 2002 pre-season (67.9 min, 95% CI: 66.8 to 68.9) and 2003 pre-season (74.5 min, 95% CI: 73.7 to 75.3) period. The 2002 and 2003 pre-season training loads were significantly lower (p<0.001) than the 2001 pre-season training loads. There were no significant differences (p>0.05) between the 2002 and 2003 pre-season training loads (table 2, fig 1).

Table 2

 Monthly training loads of sub-elite rugby league players over three consecutive pre-season preparation periods

Figure 1

 Overall training intensity, duration, and loads, and injury rates of sub-elite rugby league players over three consecutive pre-season preparation periods. Values are reported as means (95% CI). *Significantly different (p<0.001) from 2002 and 2003 seasons. †Significantly different (p<0.001) from 2002 season. #Significantly different (p<0.01) from 2003 season.

Incidence of injury

The overall injury exposure for the three pre-season periods was 1442.4 (2001), 1165.9 (2002), and 1478.9 (2003) training hours at risk. The incidence of injury was significantly higher (χ2 = 44.3, df 2, p<0.001) in the 2001 pre-season period (156.7 per 1000 training hours, 95% CI: 136.3 to 177.1) than the 2002 (94.4 per 1000 training hours, 95% CI: 76.7 to 112.0) and 2003 (78.4 per 1000 training hours, 95% CI: 64.2 to 92.7) pre-season periods. There were no significant differences in the incidence of injury between the 2002 and 2003 pre-season periods (table 3, fig 1).

Table 3

 Monthly injury rates of sub-elite rugby league players over three consecutive pre-season preparation periods

Site of injury

The majority of injuries sustained over the three seasons were to the thigh and calf, ankle and foot, and knee. There were significant differences among seasons for injuries sustained to the thorax and abdomen (χ2 = 14.6, df 2, p<0.001). The incidence of thigh and calf (χ2 = 28.2, df 2, p<0.001) and ankle and foot (χ2 = 26.9, df 2, p<0.001) injuries was significantly higher in the 2001 pre-season period than the 2002 and 2003 pre-season periods (table 4).

Table 4

 Site of training injuries sustained by sub-elite rugby league players over three consecutive pre-season preparation periods

Nature of injury

The type of injuries sustained over the three seasons is shown in table 5. The majority of injuries sustained over the three seasons were muscular strains and joint sprains. The incidence of muscular strains (χ2 = 44.6, df 2, p<0.001), joint sprains (χ2 = 17.0, df 2, p<0.001), and haematomas (χ2 = 7.1, df 2, p<0.05) was significantly higher in the 2001 pre-season period than the 2002 and 2003 pre-season periods.

Table 5

 Nature of training injuries sustained by sub-elite rugby league players over three consecutive pre-season preparation periods

Cause of injury

Overexertion was the most common cause of injury. There were significant differences among seasons for injuries sustained while being tackled (χ2 = 6.4, df 2, p<0.05). Overexertion (χ2 = 38.2, df 2, p<0.001) and overuse (χ2 = 11.1, df 2, p<0.01) injuries were more common in the 2001 pre-season period than the 2002 and 2003 pre-season periods (table 6).

Table 6

 Cause of training injuries sustained by sub-elite rugby league players over three consecutive pre-season preparation periods

Severity of injury

The majority of injuries were transient, resulting in no loss of training time. There were significantly more (χ2 = 6.0, df 2, p<0.05) severe injuries in the 2001 pre-season period (33.3 per 1000 training hours, 95% CI: 23.9 to 42.7) than the 2002 (30.0 per 1000 training hours, 95% CI: 20.0 to 40.0) and 2003 (18.9 per 1000 training hours, 95% CI: 11.9 to 25.9) pre-season periods (table 7).

Table 7

 Severity of training injuries sustained by sub-elite rugby league players over three consecutive pre-season preparation periods

Changes in physical fitness

The changes in speed, muscular power, and V˙o2max over the three pre-season periods are shown in table 8. The pre-training muscular power and V˙o2max were similar (p>0.05) across the three pre-season periods. Pre-training speed measurements (10 m, 20 m, and 40 m sprint) were significantly faster (p<0.05) in the 2003 pre-season. There were no significant (p>0.05) seasonal differences for changes in 10 m, 20 m, and 40 m speed. There were greater improvements in muscular power in the 2003 pre-season period, with a 76% probability that the improvements were of physiological significance. Each pre-season training period induced a significant increase (p<0.05) in V˙o2max. The increases in V˙o2max progressively improved across the three seasons (2001, 7.7%; 2002, 11.8%; 2003, 15.6%), with a 62–88% probability that the 2002 and 2003 pre-season improvements in V˙o2max were of greater physiological significance than the 2001 pre-season improvements in V˙o2max.

Table 8

 Pre-training and post-training measurements of muscular power, maximal aerobic power, and 10 m, 20 m, and 40 m speed of sub-elite rugby league players over three consecutive pre-season preparation periods

DISCUSSION

This study is the first to concurrently investigate training loads, changes in physical fitness, and training injury rates in rugby league players. In addition, this study is the first to implement and evaluate the effectiveness of an evidence based injury prevention program for rugby league players. The results of this study are of particular importance given that the greatest training injury rates have been shown to occur in the pre-season preparation period15,18 and that the pre-season preparation period is where the greatest changes in physical fitness can be expected to occur. These findings demonstrate that reductions in pre-season training loads reduce training injury rates in rugby league players and result in greater improvements in maximal aerobic power.

In the present study, reductions in training loads were accomplished through reductions in training duration in the 2002 pre-season and training intensity in the 2003 pre-season. Both changes to the training program elicited reductions in training injury rates. A 10.6–15.7% reduction in training loads reduced the incidence of injury by 39.8–50.0%, without compromising the pre-season improvements in physical fitness. Indeed, there was a greater relative change in V˙o2max with reduced training loads (2001, 7.7%; 2002, 11.8%; 2003, 15.6%). It is generally acknowledged that the greatest improvements in physical fitness occur in athletes with the lowest initial fitness level.27 Therefore, the greater relative changes in V˙o2max over the three pre-season periods could possibly be explained by the slightly lower pre-training V˙o2max of players in the 2002 and 2003 pre-season periods. Alternatively, the greater improvements in V˙o2max in the latter seasons may be due to the lower incidence and severity of injuries in this period, thereby allowing players to participate in a greater number of training sessions than the 2001 pre-season. The lack of significant differences in injury rates between the 2002 and 2003 pre-season periods suggests that reductions in training intensity and training duration are equally effective in reducing training injury rates. It is unclear if greater reductions in either training intensity or training duration would further reduce training injury rates. It is also unclear whether further reductions in injury rates may occur through concomitant reductions in both training duration and training intensity. Given the low incidence of injury and the large improvements in V˙o2max in the 2003 pre-season period, it is possible that further reductions in training load may reduce training injury rates but also minimise the improvements in physical fitness obtained during the pre-season preparation period. The obvious challenge for rugby league conditioning coaches is to develop training programs that provide an adequate training stimulus to enhance physical fitness, without unduly increasing the incidence of training injuries. Further studies investigating the effect of reductions in both training intensity and training duration on training injury rates and physical fitness are warranted.

In the present study, reductions in training loads were associated with reductions in lower limb injuries, muscular strains, and joint sprains. In addition, overuse injuries and injuries sustained as a result of overexertion were less common following the reductions in training loads. Overtraining is associated with a high incidence of overuse injuries28 and compromised aerobic fitness.29 Given the high incidence of lower limb overuse injuries and smaller increases in V˙o2max in the 2001 pre-season, it is likely that these players were over-trained, and that the training loads applied were greater than was tolerable for the musculoskeletal system. While the absolute training loads applied in the 2001 pre-season period were unlikely to pose a significant training stimulus for elite rugby league players or players with a greater training history, it is possible that the applied training program limited the opportunity for adequate recovery for the players training in the 2001 pre-season period. Furthermore, the higher intensity, longer duration training sessions undertaken in the 2001 pre-season period may provide some explanation for the high incidence of thigh and calf strains and greater injury severity during this period.

Age,30 playing experience,31 and previous injury30 have been shown to be risk factors for subsequent sporting injury. The finding of greater age and playing experience in the players of the 2001 season may therefore offer an explanation for the higher injury rates in this cohort. In addition, given the greater playing experience of the players in the 2001 season, it is likely that they had also sustained more injuries prior to the commencement of the study. The higher incidence of injury during the 2001 pre-season preparation period may reflect a greater risk of injury due to previous injuries sustained while participating in rugby league.

Previous studies of collision sports (Australian football) have shown that a low rainfall is a significant predictor of injury, and that injury rates are reduced when rainfall is high.25 The present study found a similar rainfall in the 2001 and 2003 pre-season periods, but a considerably lower rainfall in the 2002 pre-season period. In contrast to previous findings,25 the incidence of injury was significantly reduced when rainfall was lowest (2002 pre-season period), and remained low when rainfall returned to similar values to those obtained during the 2001 pre-season. Given the similar rainfall between the 2001 and 2003 pre-season periods, the lower rainfall in the 2002 pre-season period, and the marked differences in injury rates, it is unlikely that rainfall influenced the incidence of injury in the present study. The reason for the different findings between the present and previous studies25 is unclear but may be due to the duration of the studies, with the present investigation spanning a 4 month pre-season period, and previous investigations spanning an entire playing season. Further studies are required to determine the relationship between environmental conditions (particularly rainfall) and injury rates in rugby league over an entire playing season.

The present findings provide the expected changes in physical fitness from a 4 month pre-season conditioning program in sub-elite rugby league players. These findings demonstrate that sub-elite rugby league players undertaking a progressively overloaded training program, performing two sessions per week, may expect a 7.5–15.9% increase in aerobic fitness, and stable 10 m, 20 m, and 40 m speed. The improvements in V˙o2max are higher than the 1.9–3.0% improvement in V˙o2max previously reported for elite rugby union players,32,33 and similar to that previously reported (10.2%) for elite soccer players34 following a similar period of pre-season conditioning. In addition, the 5.7% improvement in vertical jump scores during the 2003 pre-season period was well above the typical error of measurement of 3.3%, therefore offering players a physiologically significant increase in muscular power. As only two training sessions were performed each week, with training sessions generally no longer than 90 min in duration, it was difficult to train all physiological fitness parameters effectively and also enhance skill. It is unclear if greater improvements in the physiological capacities of players would have occurred if a larger amount of training time was devoted to the development of specific performance parameters (for example, aerobic fitness, muscular power, and speed). Equally, it is unclear if further improvements in V˙o2max, muscular power, and speed would have occurred during the competitive phase of the season, when players were required to contend with injuries and residual fatigue.32 Future studies documenting the training loads and seasonal changes in the physiological performance characteristics of rugby league players are clearly required.

What is already known on this topic

The majority of rugby league training injuries occur in the pre-season preparation period when training loads are greatest.

What this study adds

This study found that reductions in pre-season training loads reduce training injury rates in rugby league players and result in greater improvements in maximal aerobic power.

Due to limited resources (for example, video analysis, match statistics), it was not possible to analyse playing performance (for example, distance covered during a match, number of involvements with and without the ball, average work intensity during a match, etc) in the present study. While playing performance was not recorded, the win-loss ratio declined over the three seasons and coincided with the reductions in training loads. These results suggest that reductions in training loads may be associated with reductions in playing performance. However, the playing experience of the subjects also declined over the 3 year period (table 1). It has previously been shown that playing experience is a significant predictor of successful rugby league performance.3 Therefore, the lower playing experience of subjects in the 2002 and 2003 seasons may explain the reduction in playing performance (as evidenced from the win-loss ratio) during this period. Clearly, further studies investigating the effect of reductions in training loads on playing performance are warranted.

In summary, the present study investigated if reductions in pre-season training loads reduced the incidence of training injuries in rugby league players and determined if the reductions in training loads compromised the improvements in physical fitness obtained during the pre-season preparation period. The results of this study demonstrate that reductions in pre-season training loads reduce training injury rates in rugby league players and result in greater improvements in maximal aerobic power. Further studies investigating the influence of reductions in training loads on the playing performance of rugby league players are required.

REFERENCES

View Abstract

Footnotes

  • Conflict of interest: none declared.