Skip to main content

SYSTEMATIC REVIEW article

Front. Physiol., 10 April 2018
Sec. Exercise Physiology

Tests for the Assessment of Sport-Specific Performance in Olympic Combat Sports: A Systematic Review With Practical Recommendations

  • 1Division of Training and Movement Sciences, Research Focus Cognition Sciences, University of Potsdam, Potsdam, Germany
  • 2Research Unit (UR17JS01) “Sport Performance & Health” Higher Institute of Sport and Physical Education of Ksar Said, Tunis, Tunisia
  • 3Department of Human Movement and Sport Sciences, University of Rome “Foro Italico,” Rome, Italy
  • 4Martial Arts and Combat Sports Research Group, School of Physical Education and Sport, Universidade de São Paulo, São Paulo, Brazil
  • 5Laboratoire Psychologie de la Perception, UMR Centre National de la Recherche Scientifique 8242, Paris, France
  • 6Cesam, EA 4260, Université de Caen Basse-Normandie, UNICAEN, Paris, France

The regular monitoring of physical fitness and sport-specific performance is important in elite sports to increase the likelihood of success in competition. This study aimed to systematically review and to critically appraise the methodological quality, validation data, and feasibility of the sport-specific performance assessment in Olympic combat sports like amateur boxing, fencing, judo, karate, taekwondo, and wrestling. A systematic search was conducted in the electronic databases PubMed, Google-Scholar, and Science-Direct up to October 2017. Studies in combat sports were included that reported validation data (e.g., reliability, validity, sensitivity) of sport-specific tests. Overall, 39 studies were eligible for inclusion in this review. The majority of studies (74%) contained sample sizes <30 subjects. Nearly, 1/3 of the reviewed studies lacked a sufficient description (e.g., anthropometrics, age, expertise level) of the included participants. Seventy-two percent of studies did not sufficiently report inclusion/exclusion criteria of their participants. In 62% of the included studies, the description and/or inclusion of a familiarization session (s) was either incomplete or not existent. Sixty-percent of studies did not report any details about the stability of testing conditions. Approximately half of the studies examined reliability measures of the included sport-specific tests (intraclass correlation coefficient [ICC] = 0.43–1.00). Content validity was addressed in all included studies, criterion validity (only the concurrent aspect of it) in approximately half of the studies with correlation coefficients ranging from r = −0.41 to 0.90. Construct validity was reported in 31% of the included studies and predictive validity in only one. Test sensitivity was addressed in 13% of the included studies. The majority of studies (64%) ignored and/or provided incomplete information on test feasibility and methodological limitations of the sport-specific test. In 28% of the included studies, insufficient information or a complete lack of information was provided in the respective field of the test application. Several methodological gaps exist in studies that used sport-specific performance tests in Olympic combat sports. Additional research should adopt more rigorous validation procedures in the application and description of sport-specific performance tests in Olympic combat sports.

Introduction

Amateur boxing, fencing, karate, judo, taekwondo, and wrestling represent popular combat sports. These combat sports are practiced in the whole world and constitute an important part of the Summer Olympic programme (International Olympic Committee, 2017a). Wrestling and fencing were already part of the first modern Olympic Games in 1896 for males. Females were included in 1924 for fencing and in 2004 for wrestling. In 1904, male amateur boxing was included in the official program of the Summer Olympic Games. It lasted until 2012, until female amateur boxing became part of the Olympic program. In 1964, judo was included in the Olympic program for males and in 1992 for females. Taekwondo was recognized as an Olympic sport in 2000 for both sexes and karate will be introduced for both sexes in the 2020 Olympic Games. In this regard and with reference to the growing interest in Olympic combat sports, it is important to advance scientific knowledge in performance testing to design specifically tailored training protocols and periodization models and to increase the likelihood of success in competition (Bridge et al., 2014; Chaabene et al., 2017a).

The main purposes of sport-specific testing can comprise talent identification and development of young athletes as well as the identification of strengths and weaknesses in young and elite athletes to be used for training purposes (Tabben et al., 2014; Chaabane and Negra, 2015). In addition, there is a consensus in the scientific literature on the importance of assessing physical and physiological qualities to optimize sport performance (Franchini et al., 2011a; Bridge et al., 2014; Chaabène et al., 2015c; Chaabene et al., 2017b) especially for those characterized by complex technical/tactical and physical/physiological demands like striking (e.g., karate, taekwondo, and amateur boxing; Chaabène et al., 2012b, 2015c; Bridge et al., 2014), grappling (e.g., judo and wrestling; Franchini et al., 2011a; Chaabene et al., 2017a), and weapon-based combat sports (e.g., fencing; Roi and Bianchedi, 2008).

However, prior to the design of a test protocol for sport-specific performance assessment, it is recommended to conduct a systematic needs analysis to identify the above-mentioned demands of the specific sport (Kraemer et al., 2012). More specifically, in the context of a need analysis, the metabolic, biomechanical, and injuries profile of the sport could be explored (Kraemer et al., 2012). With the systematically derived information on sport-specific demands from the needs analysis, adequate sport-specific performance tests can be designed and implemented into training practice. Information from these tests allows to identify strengths and weaknesses of athletes and to monitor how athletes' performance developed over time. These individualized performance profiles of athletes can be used for the planning of training protocols and periodization models. In Olympic combat sports, growing number of researchers have turned their attention to the development of valid sport-specific test protocols that are specifically tailored to the physical, physiological, technical and tactical demands of the respective sport discipline (Santos et al., 2010; Chaabène et al., 2012c; Tabben et al., 2014; Sant'Ana et al., 2017).

Even though there is a well-accepted advantage of sport-specific performance testing over the application of general physical fitness tests, there is no study available that systematically reviewed the methodological quality (e.g., sample size, inclusion/exclusion criteria, stability of testing conditions), validation data (i.e., reliability, validity, sensitivity), and feasibility (i.e., practicability) of the existing sport-specific tests related to Olympic combat sports. In fact, the majority of the available literature focused on the assessment of physical and physiological attributes of Olympic combat sport athletes in sports like amateur boxing (Chaabène et al., 2015c), fencing (Roi and Bianchedi, 2008), judo (Franchini et al., 2011a), karate (Chaabène et al., 2012b, 2015a), taekwondo (Bridge et al., 2014), and wrestling (Chaabene et al., 2017a). However, these tests assess general physical fitness qualities but not sport-specific performance. To the authors' knowledge, previous systematic reviews (Robertson et al., 2014; Hulteen et al., 2015) critically appraised the methodological quality and feasibility of performance tests in individual (e.g., golf, tennis, rock climbing) and team sports (e.g., football, rugby, volleyball) but not in Olympic combat sports. Therefore, the purpose of this study was to systematically review the available literature and to critically analyze the methodological quality, validation data, and feasibility of sport-specific tests in Olympic combat sports.

Methods

The experimental approach comprehended five-steps (Khan et al., 2003): Step 1: Framing questions for the review; Step 2: Identification of relevant works; Step 3: Assessment of the quality of studies; Step 4: Summary the evidence; and Step 5: Interpretation of the findings.

Step 1: Framing Questions for the Review

The research question focused on sport-specific testing in Olympic combat sports (e.g., boxing, fencing, judo, karate, taekwondo, and wrestling). A Boolean search strategy was applied using the operators AND, OR. According to the main topic of the present study, the a-priori-specified inclusion criteria encompassed the following search syntax: [(“combat sport*” OR karate OR taekwondo OR “amateur boxing” OR judo OR wrestling OR fencing) AND (reliability OR validity OR sensitivity) AND (“physical fitness” OR “physiological characteristic*” OR “physical activity” OR “fitness test*” OR “motor assessment” OR “technical skill*” OR “gold standard”)].

Step 2: Identification of Relevant Works and Data Extraction

The present systematic review of the published literature was conducted based on the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) guidelines (Moher et al., 2009). A comprehensive literature search of original manuscripts investigating was systematically performed on PubMed (MEDLINE), Google Scholar, and Science Direct. The search was limited to manuscripts published up to October 2017. In this study, the criteria for the inclusion of retrieved articles were: (i) written in English, (ii) published in peer-reviewed journals, (iii) focused on either on amateur boxing, fencing, judo, karate, taekwondo, wrestling, or a combination of these combat sports (iv) evaluate one aspect of the physical fitness and/or physiological characteristic through sport-specific testing, and (v) report at least one aspect of either reliability, validity, or sensitivity related to the applied test protocol. To allow the assessment of the methodological quality, only full-text sources were included, whereas abstracts and conference papers from annual meetings were not considered in the analysis. The first author (HC) coded the studies according to the selection criteria and eliminated duplicates. Relevant articles identified through the searching process were independently evaluated and assessed by two reviewers (e.g., HC and YN) who screened the titles, the abstracts, and the full texts to reach the final decision on the study inclusion or exclusion. In case of uncertainty or disagreement, a third expert was consulted. Additionally, the snowballing technique was applied to the reference lists of retrieved full-text articles to identify additional articles that were not included in the initial electronic search (Figure 1).

FIGURE 1
www.frontiersin.org

Figure 1. Flow of information through the different phases of the systematic review.

For each selected study included in the final list of scientific contributions considered eligible for the detailed examination, data were extracted and examined by two independent reviewers (HC and YN) who conducted data extraction following a predefined template. The template included sample size, demographic information including sex, age, and training background/expertise, type of combat sports, sport-specific test's name, objective, and duration, and validation data (e.g., reliability, validity, and sensitivity). After completion of data extraction, the two independent reviewers cross-checked the data to confirm their accuracy. Any conflicting results between reviewers resulted in a re-evaluation of the paper in question until a consensus was reached. It is worth noting that information related to the type of reliability (e.g., test-retest reliability and inter/intra-rater reliability; Currell and Jeukendrup, 2008; Robertson et al., 2014), validity (e.g., content or logical, criterion [concurrent/predictive], and construct [discriminative/convergent]) (Currell and Jeukendrup, 2008; Robertson et al., 2014), and sensitivity (Currell and Jeukendrup, 2008; Impellizzeri and Marcora, 2009) with the used corresponding statistical tools were specified.

Table 1 provides details of items related to the sport-specific tests' characteristics and feasibility. In addition, the most common statistical analysis approaches used to assess reliability were considered (Hopkins, 2000; Currell and Jeukendrup, 2008; Impellizzeri and Marcora, 2009), which encompassed: coefficient of variation (CV%), intra-class correlation coefficient (ICC), typical error of measurement (TEM), correlation coefficients (r), and 95% limits of agreement (LOA). Additionally, details related to the validity of the sport-specific testing were retrieved. Test sensitivity was assessed through comparing the smallest worthwhile change (SWC) and the typical error of measurement (TEM). Furthermore, the minimal detectable change (MDC95%), was also included (Hopkins, 2002).

TABLE 1
www.frontiersin.org

Table 1. Details related to sport-specific tests' characteristics.

Step 3: Assessment of the Quality of Studies

Based upon six criteria adapted from a risk-of-bias evaluation in a previously published review of tests examining sport-related skill outcomes (Robertson et al., 2014), two authors (HC and YN) carefully reviewed all eligible articles for quality appraisal (Table 3). These criteria included: (i) sample size, i.e., the number of participants included to establish validity/reliability/sensitivity of the sport-specific test; (ii) details related to study participants (e.g., sex, age, sport background/expertise, and anthropometric details); (iii) presence of clearly established inclusion/exclusion criteria; (iv) presence of familiarization session(s) prior to sport-specific testing or not; (v) presence of clearly established interval between test-retest assessments; and (vi) stability of testing conditions (i.e., whether testing equipment and environmental conditions remained stable between sessions or not) as well as participants between sessions.

Step 4: Summary the Evidence

From each included study, the following information was extracted: author(s), year of publication, country, aim of the study, research design: Then, detailed tables reporting major characteristics of the selected studies were created.

Step 5: Interpretation of the Findings

A synthesis of major findings reported in the included articles was submitted to a thematic analysis deemed relevant for generating inferences.

Results

The preliminary systematic search resulted in 4,755 hits. After careful examination of titles and abstracts, 428 articles remained and were reviewed for eligibility. Full texts of these 428 articles were screened in regards of the previously defined inclusion/exclusion criteria. Finally, 39 articles were eligible to be included in this systematic review (Figure 1). Table 2 illustrates the main characteristics of the 39 eligible scientific contributions in terms of the applied sport-specific test, the tested physical and/or physiological attributes, the measured outcomes, and the number of female and male participants and their athletic background/expertise. Publication years ranged from 1997 (Utter et al., 1997) to 2017 (Araujo et al., 2017; Chaabene et al., 2017b; Sant'Ana et al., 2017), with an increasing trend starting from 2007 (Figure 2). Judo was represented in 15 articles (39%), whereas the relative picture for taekwondo, karate, fencing, wrestling, and amateur boxing was 8 (20%), 7 (18%), 4 (10%), 3 on (8%), and 2 (5%), respectively. The number of participants ranged from 5 to 219, with the majority of studies (74%) including <30 participants, whereas the proportion of studies including 30–49, 50–99, and >100 participants was 5, 13, and 3%, respectively. It is noteworthy that one study (Shiyan, 2011) did not include any information about participants' number. Detailed information on participants' characteristics were reported in 67% of the eligible studies with 31% providing a partial description, and another 2% lacking information regarding the recruited participants (Table 2). The proportion of studies including athletes competing at national or international level was 41%, that including mixed samples of national/international and regional/sub-elite level athletes was 33%, and that recruiting club/regional level participants was 26%. The sex representation of the participants was 61% for male athletes, 5% for female athletes, and 28% for both sexes. No information on the sex of the participants was present for 6% of the studies.

TABLE 2
www.frontiersin.org

Table 2. Characteristics of the eligible studies.

FIGURE 2
www.frontiersin.org

Figure 2. Frequency of Occurrence (%) of publication years of the scientific contributions included in the systematic literature review.

Methodological Quality of the Eligible Studies

The methodological characteristics of the included studies are shown in Table 3. Whilst only 28% of the included studies provided detailed inclusion/exclusion criteria, 26% provided a partial description and 46% did not include any information. Only 38% of the studies scheduled one or more familiarization sessions for their participants and 2% provided partial details on this relevant aspect, whereas the majority of the studies (60%) presented a lack of information. Studies focusing on reliability (61%) included intervals between experimental trials ranging from 1 day to 11 weeks, with 57% of them adopting a 1-week interval. However, information on this issue was lacking in 39% of studies. Only 28% of the studies provided detailed information on stability of testing conditions, with 59% of them not presenting any information and 13% providing partial information.

TABLE 3
www.frontiersin.org

Table 3. Methodological quality of eligible studies.

Reliability

Table 4 illustrates detailed information on validation data of the respective sport-specific tests. From the 39 included studies, 51% reported reliability data of the sport-specific test. Test-retest reliability was examined in 85% of the studies that conducted reliability analysis, with 15% dealing with inter/intra-rater reliability. The ICC (range 0.43–1) was used in 80% of the studies and constituted, therefore, the most frequently applied statistical approach to assess reliability in sport-specific tests of Olympic combat sports. The SEM was used in 30% of the identified studies, the CV in 5%, the 95% LoA in 15%, and correlation coefficients in 5% of the studies. Other statistical approaches (i.e., paired sample t-tests and Wilcoxon signed rank tests) were used to establish reliability in 15% of the studies. Only 40% of the studies applied mixed statistical approaches to examine relative and absolute reliability as recommended by previous research (Atkinson and Nevill, 1998) and the greater part of them (60%) adopted only one statistical approach (most often ICC [67%]).

TABLE 4
www.frontiersin.org

Table 4. Sport specific tests characteristics.

Validity and Sensitivity

All reviewed studies presented at least one aspect of test validity. Of note, content validity was addressed in all identified studies. Criterion validity was determined in 54% of the eligible studies, with 95% addressing concurrent validity (r = −0.41 to 0.90) and 5% predictive validity. From the studies that addressed concurrent validity, 60% applied correlation coefficients only, whereas 10% used mixed correlation coefficients with other methods (e.g., 95% LOA, regression analysis), and 30% applied other approaches (e.g., Wilcoxon signed rank test, paired sample T-test, 95% LOA). Construct validity was examined in 31% of the identified 39 studies, with the discriminative side of it (i.e., the ability of the sport-specific test to differentiate performance according to expertise level) being the most important aspect in all studies. This was realized by computing receiver operator characteristic (ROC) analyses, independent sample t-tests, and two-way repeated measures ANOVA. Only 23% of the identified studies addressed content validity, 77% examined mixed aspects (e.g., content with criterion validity, content with construct validity or content with the criterion and construct validity). One of the three sport-specific testing aspects (e.g., content, criterion, and construct validity) was investigated in only 8% of the studies. The sensitivity of sport-specific testing was investigated in 13% of the reviewed studies. These studies mainly calculated the SWC and compared it with SEM (Chaabène et al., 2012a; Chaabene et al., 2017b; Tabben et al., 2014) and one study (da Silva Santos and Franchini, 2016) used data recorded after a 9-week training period to appraise sensitivity of the respective sport-specific test. The MDC95% of the sport-specific test was addressed in 8% of the reviewed studies.

Utility and Limitations

Feasibility and methodological limitations of the sport-specific tests were sufficiently explored in 36% of the included studies, with 51% providing partial details and 13% ignoring this aspect. Information related to the expected use and context of the test was adequately pointed out in 72% of the included studies. Eighteen percent reported limited information on this aspect and 10% ignored this relevant sport-specific aspect. Sport-specific test duration was described in detail in 74% of the included studies and the remaining 26% did not report any information on this issue.

Discussion

The main goal of this study was to examine the methodological quality, validation data, and feasibility of sport-specific tests in Olympic combat sports. This is the first study detailing the different methodological approaches adopted so far with sport-specific tests in Olympic combat sports. Results of this study highlighted: (1) emerging academics conversation on sport-specific tests in Olympic combat sports; (2) a disparity in the gender representation of participants; and (3) several methodological gaps in the study of sport-specific testing in Olympic combat sports.

Since 2006 a substantial increase in publication activity has been observed, coherently with the quest of sport-specific testing procedures to evaluate Olympic combat sport athletes. At present, research in sport-specific testing of Olympic combat sports could be considered entering its intermediate stage (Edmondson and McManus, 2007), being characterized by not fully established theories and several methodological shortcomings. In particular, a lack of a “gold standard” technique to assess sport-specific outcomes in peculiar combat sport contexts and valid and reliable tools suited to large-scale assessments limits the generalizability of findings. Furthermore, in considering that ~40% of the eligible studies focused on judo, the need to develop valid sport-specific tests for athletes practicing other Olympic combat sports emerged.

Overall, a major challenge for the interpretation of sport-specific test data to be used for training periodization is due to methodological limits. Even though researchers have attempted to develop and validate sport-specific tests in Olympic combat sports, future studies should carefully address methodological aspects. More specifically, further research should focus on i) tests that accurately reflect athletes' sport-specific performance strengths and weaknesses and ii) present good level of predictive validity. Particularly, special attention should be directed toward the recruitment of a wide range of athletes, detailing clear inclusion/exclusion criteria of participants, and presenting sufficient description of their characteristics such as anthropometrics, age, and expertise level. This issue is crucial to guarantee the test-specificity for different populations of athletes, which allows coaches to programme sound individualized training plans. Furthermore, clear and comprehensive information on test procedures is needed so the protocol can be easily reproduced (Morrow et al., 2015). Another aspect to be considered is the provision of information on test-retest reliability, including test-retest intervals, intra- and inter-rater reliability, and the stability of testing conditions, which could determine problems in the interpretation of results (Atkinson and Nevill, 1998; Morrow et al., 2015).

In general, validity, reliability and sensitivity are basic criteria for a test able to assess sport performances. When examining validation data, approximately half of the included studies examined reliability of the sport-specific test using test-retest as the most frequently applied reliability aspect and ICCs were most often computed (80%). Content validity was addressed in all identified studies. Criterion validity and more specifically the concurrent side was assessed in approximately half of the studies, while construct validity received less attention (31% of the studies). Of note, predictive validity was surprisingly neglected. In fact, only one study that examined this test characteristic (Lidor et al., 2005) has been identified. Additionally, few studies examined test sensitivity (13%). Feasibility and methodological limitations were partially reported and/or ignored in the majority of the reviewed (64%) studies. Detailed information related to the expected use and context of the protocol were either partially reported or ignored in 28% of the studies.

Methodological Quality of the Included Studies

One major point related to the methodological quality is the limited sample size recruited in the majority of the reviewed studies. It is consensual that sample size is the most critical aspect decoding study's outcome quality and applicability (Hopkins et al., 2009). In this contest, 5 studies included between 50 and 99 participants (Sterkowicz and Franchini, 2001; Sertić et al., 2011; Weichenberger et al., 2012; Turner et al., 2016), one study included more than 100 participants (Sterkowicz and Franchini, 2009), and one study did not provide participants' number (Shiyan, 2011). This observation seems to be due to the limited number of coaches agreeing their athletes to be involved in such studies. One more issue that may prevent and/or question sport-specific tests to be applied with other population, for instance, amateur and beginner practitioners, is the recruitment of national/international level athletes in most of the studies. Compared with males, females were recruited in two studies (Santos et al., 2012; Tavra et al., 2016), with 28% of the studies recruiting combat sport athletes of both sexes (Tsolakis and Vagenas, 2010; Obminski et al., 2011; Weichenberger et al., 2012; Del Vecchio et al., 2014; Tabben et al., 2014; Chaabène et al., 2015b; Chaabene et al., 2017b; Oliveira et al., 2015; da Silva Santos and Franchini, 2016; Morales et al., 2016; Turner et al., 2016). This seems to be mainly due to the limited interest and/or opportunity of females in combat sports, as only recently female competitions were included in Olympic boxing and wrestling, for instance. Also, cultural constraints to female participation in combat sports (Miarka et al., 2011) may determine the gender-related discrepancies in the sport sciences literature, which does not mirror the increased participation of women in the last editions of the Olympic Games (International Olympic Committee, 2017b). Therefore, the sports scholars are urged to intensify their efforts to bridge this imbalance between women's sport participation and scientific information on this specific population.

Details related to the recruited participants were either partially reported or ignored in 33% of the studies (Utter et al., 1997; Franchini et al., 1998, 2005; Sterkowicz and Franchini, 2009; Obminski et al., 2011; Sertić et al., 2011; Shiyan, 2011; Weichenberger et al., 2012; Bottoms et al., 2013; Oliveira et al., 2015; Sogabe et al., 2015; da Silva Santos and Franchini, 2016; Rocha et al., 2016). This issue markedly affects the quality of the study and prevents the sport-specific test of being replicated and used. There is a lack (26% of the studies) and most often absence (46% of the studies) of any inclusion/exclusion criteria and only 28% of the studies sufficiently detailed this aspect. Therefore, future investigations are encouraged to consider clarifying this important research aspect. Despite their relevance in reducing measurement error, mainly systematic bias in terms of learning effects (Atkinson and Nevill, 1998), familiarization sessions were considered in only 38% of the studies with the most of them (60% of the studies) neglected this aspect and 2% provided limited details. This may increase sport-specific measurement bias and affect, thereafter, the accuracy of the test. The most adopted test-retest interval in 61% of the reviewed studies was 1 week. It should be noted that the test-retest interval should not be too short to avoid insufficient recovery between tests (Atkinson and Nevill, 1998) or too long to avoid being affected by participant's skill enhancement between the test and retest (Robertson et al., 2014). However, the exact test-retest interval is mainly dependent on the sport-specific test's characteristics in terms of complexity, duration, and type of effort required. Regarding the stability of testing conditions, most of the studies did not provide any (59%) or provided partial (13%) details. Again, this may affect the quality and accuracy of the sport-specific outcomes as different environmental conditions, for instance, may considerably influence testing results (Hachana et al., 2012).

Reliability

Reliability is the ability of the testing protocol to provide similar outcomes from day to day when no intervention is used (Atkinson and Nevill, 1998). It is an important testing aspect as it provides indications about the biological as well as technical variation of the protocol (Bagger et al., 2003). From the three main aspects of reliability (i.e., test-retest, intra-, and inter-rater reliability, for in-depth details see Table 1), test-retest reliability represents the most studied sport-specific property (85% of studies that examined reliability) compared with intra/inter-rater reliability (15% of the studies) (Table 4). To effectively establish reliability, previous studies recommended determining both types of it i.e., relative and absolute reliability (Atkinson and Nevill, 1998; Weir, 2005; Impellizzeri and Marcora, 2009). To do so, a mixed statistical approach could be used, for instance, ICC which is indicative of the relative reliability of a test and SEM which is indicative of its absolute aspect (Atkinson and Nevill, 1998; Weir, 2005). Results of the current review showed that only few studies applied a mixed approach to examine both relative and absolute reliability of their sport-specific tests (40% of the studies that examined reliability). As this may constitute a limitation, upcoming investigations need to establish both types of reliability. On the other hand, a number of studies used other statistical approaches such as paired sample t-test (10% of the studies that addressed reliability). However, such an approach has been criticized in a previous review (Atkinson and Nevill, 1998) in the way that it does not provide any indication of random variation between tests. Additionally, Bland and Altman (1995) recommended paying attention to the interpretation of paired t-test results of reliability mainly because the detection of a significant difference is actually dependent on the amount of random variation between tests. Overall, to accurately establish sport-specific test's reliability, it is recommended to calculate both relative and absolute reliability by adopting appropriate statistical approaches. In addition, the other reliability aspects (i.e., inter/intra-rater) need to be investigated in conjunction with test-retest reliability. In that manner, a clear and accurate overview about the sport-specific test's reliability can be drawn. In fact, ensuring of test's reliability at first will enable the researcher to move on to check aspects related to validity and sensitivity. Otherwise, the testing protocol will be judged as non-valid. In this regard, Atkinson and Nevill (1998) argued that a measurement tool will never be valid if it provides inconsistent outcomes from repeated measurements.

Validity and Sensitivity

Content validity was established in all the reviewed studies. This is obvious since one of the current study's inclusion criteria is to deal with a sport-specific testing. Content validity was generally assumed (i.e., in 98% of the studies) by mainly referring to the specific literature and appraisal of the actual competition/combat requirements. However, only one study (Tabben et al., 2014) established this test's quality through a mixing of previous consultation with combat sports practitioners, coaches, sports scientists, and a review of the literature and competition requirements. Criterion validity was addressed in approximately half of the reviewed studies. The major part of these studies (95%) considered concurrent validity. Concurrent validity was generally studied by associating the sport-specific testing's outcome with a gold standard protocol (e.g., treadmill running test, cycle ergometer test). These gold standard tests are based on actions and thereafter involve muscle groups that are not combat sport specific, which may affect findings related to test's property (e.g., results reflective of a poor concurrent validity when the test reflect the true sport-related effort or findings indicative of good concurrent validity when the test did not reflect the true sport-related effort). From the whole eligible studies considered in this review, only one study addressed predictive validity (Lidor et al., 2005). This is particularly surprising in view of the critical importance of such a testing's property for coaches, strength and conditioning professionals, and combat sports athletes (Currell and Jeukendrup, 2008; Robertson et al., 2014). Therefore, future studies are encouraged to establish this important sport-specific testing's aspect.

Compared with criterion validity, construct validity received less attention in the literature (31% of the studies). Of note, only the discriminative side of construct validity was addressed by mainly comparing combat sports practitioners with a different competitive level and/or background (e.g., international vs. national level, elite vs. sub-elite) (Franchini et al., 1998, 2005, 2011b; Smith et al., 2000; Sterkowicz and Franchini, 2001; Tsolakis and Vagenas, 2010; Chaabène et al., 2012c; Chaabene et al., 2017b; Weichenberger et al., 2012; Del Vecchio et al., 2014; Chen et al., 2015; Tavra et al., 2016). To do so, the main statistical approach used were ROC analysis (Chaabène et al., 2012c; Chaabene et al., 2017b), independent sample t-test (Tsolakis and Vagenas, 2010; Weichenberger et al., 2012; Del Vecchio et al., 2014; Chen et al., 2015; Tavra et al., 2016), two-way repeated measures ANOVA (Franchini et al., 1998, 2011b; Smith et al., 2000; Sterkowicz and Franchini, 2001), and ANCOVA (Franchini et al., 2005). Nevertheless, as ROC analysis seems to be the more appropriate statistical approach to study discriminative ability of a test (Chaabène et al., 2012c; Chaabene et al., 2017b; Castagna et al., 2014), future investigations are recommended to use this approach. Regarding convergent validity, it was not studied in any of the reviewed studies. This seems to be due to the fact that creating a new sport-specific test is mainly due to a gap in the literature so there is no previous protocol to compare with the new one (Streiner and Norman, 2005). It is noteworthy that 8% of the studies (Weichenberger et al., 2012; Del Vecchio et al., 2014; Chaabene et al., 2017b) addressed, at least, one aspect of the three sport-specific properties (i.e., content, criterion, and construct validity). This observation may constitute another gap in the literature because, to be considered valid and applicable, a sport-specific test should cover the whole validity aspects (i.e., content, criterion, and construct validity). Thereafter, a particular focus in the future investigations should be given to examining all validity properties of sport-specific performance testing in Olympic combat sports.

Another important property related to sport-specific testing is the sensitivity (Currell and Jeukendrup, 2008; Impellizzeri and Marcora, 2009). Findings of the current review showed that only 4 studies examined this aspect (Chaabène et al., 2012c; Chaabene et al., 2017b; Tabben et al., 2014; da Silva Santos and Franchini, 2016). Additionally, despite its importance from a practical point of view, the minimal detectable change was investigated in only 3 studies (Chaabène et al., 2012a; Chaabene et al., 2017b; Tabben et al., 2014). Therefore, more research dealing with these two determinant aspects are required.

Utility and Limitations

In reviewing studies that aimed to validate sport-specific tests, thorough details about the applicability (i.e., whether it is easy to administer and scored) and the limits of the test in question were expected. However, most of the selected studies (64%) either partially detailed or ignored this valuable aspect. Details related to sport-specific test background were either partially or even ignored in 28% of the studies (Franchini et al., 1998, 2005; Smith et al., 2000; Tsolakis and Vagenas, 2010; Obminski et al., 2011; Santos et al., 2011; Sertić et al., 2011; Shiyan, 2011; Oliveira et al., 2015; Sogabe et al., 2015; Tavra et al., 2016). Therefore, future investigations should pay attention to these central sport-specific tests' aspects.

Limitations

Because of the variety of statistical approaches used to assess sport-specific measurements properties, it was not possible to perform any meta-analysis (Robertson et al., 2014). Additionally, compared with other sporting activities such as team sports, scientific contributions on combat sports in indexed journals are limited, with studies mainly published in non-indexed journals (i.e., gray literature) or remain even unpublished. Therefore, the stringent search approach adopted in this review has neglected information available to coaches in specific technical magazines and websites.

Conclusions and Future Recommendations

Establishing valid sport-specific tests that assess the actual physical fitness and/or physiological attributes of Olympic combat sports practitioners still one of the major concerns for sports sciences scholars. After reviewing 39 studies in different Olympic combat sports disciplines (e.g., karate, taekwondo, amateur boxing, judo, wrestling, and fencing), several methodological gaps have been pointed-out. These limits may prevent sport-specific testing from being widely used. These limitations are mainly related to the small sample size, backgrounds of participants, being elite level in most of the studies, sex (mainly males), lack of details about the inclusion/exclusion criteria in most of the studies, lack of familiarization session prior to testing, and paucity of details about stability of testing conditions. Additionally, both types of reliability (e.g., relative and absolute) have rarely been addressed in the reviewed studies and the available results showed reliability levels ranging from poor to excellent. Moreover, despite its critical importance, predictive validity was reported in only one study. Similarly, compared with criterion validity, construct validity received less attention by researchers. Studies addressing, at least, one aspect of the three main validity properties are limited. All these concerns may limit the applicability, generality, and accuracy of outcomes of sport-specific testing in Olympic combat sports. Additional research should adopt more strict validation procedures by addressing reliability, validity, and sensitivity in the application and description of sport-specific performance tests in Olympic combat sports. Additionally, predictive validity should receive more attention in future research.

Author Contributions

HC: Worked on study design, data collection, data analysis, and manuscript preparation. YN: Worked on data collection, data analysis, and manuscript preparation. RB: Assisted on data collection and analysis and worked on manuscript preparation. LC: Assisted in study design and worked on manuscript preparation. EF: Worked on manuscript preparation. OP: Worked on manuscript preparation. HH: Worked on manuscript preparation. UG: Data analysis and manuscript preparation.

Conflict of Interest Statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

This study is supported by the open access Publishing Fund of University of Potsdam, Germany.

References

Ali, A., Foskett, A., and Gant, N. (2008). Validation of a soccer skill test for use with females. Int. J. Sports Med. 29, 917–921. doi: 10.1055/s-2008-1038622.

PubMed Abstract | CrossRef Full Text | Google Scholar

Almansba, R., Sterkowicz, S., Sterkowicz-Przybycien, K., and Comtois, A. (2012). Reliability of the Uchikomi Fitness Test: A Pilot study. Reliability of the Uchikomi Fitness Test. Sci. Sports 27, 115–118. doi: 10.1016/j.scispo.2011.09.001

CrossRef Full Text | Google Scholar

Araujo, M. P., Nobrega, A. C. L., Espinosa, G., Hausen, M. R., Castro, R. R. T., Soares, P. P., et al. (2017). Proposal of a new specific cardiopulmonary exercise test for taekwondo athletes. J. Strength Cond. Res. 31, 1525–1535. doi: 10.1519/jsc.0000000000001312

PubMed Abstract | CrossRef Full Text | Google Scholar

Atkinson, G., and Nevill, A. M. (1998). Statistical methods for assessing measurement error (reliability) in variables relevant to sports medicine. Sports Med. 26, 217–238.

PubMed Abstract | Google Scholar

Azevedo, P. H., Drigo, A. J., Carvalho, M. C., Oliveira, J. C., Nunes, J. E., Baldissera, V., et al. (2007). Determination of judo endurance performance using the uchi - komi technique and an adapted lactate minimum test. J. Sports Sci. Med. 6, 10–14.

PubMed Abstract | Google Scholar

Bagger, M., Petersen, P. H., and Pedersen, P. K. (2003). Biological variation in variables associated with exercise training. Int. J. Sports Med. 24, 433–440. doi: 10.1055/s-2003-41180.

PubMed Abstract | CrossRef Full Text | Google Scholar

Barrow, H. M., McGee, R., and Tritschler, K. A. (1989). Practical Measurement in Physical Education and Sport. Philadelphia, PA: Lea & Febiger.

Baumgartner, T. A., and Jackson, A. S. (1998). Measurement for Evaluation in Physical Education and Exercise Science. Boston, MA: WCB/McGraw-Hill.

Bland, J. M., and Altman, D. G. (1995). Comparing two methods of clinical measurement: a personal history. Int. J. Epidemiol. 24, S7–S14.

PubMed Abstract | Google Scholar

Bottoms, L., Sinclair, J., Rome, P., Gregory, K., and Price, M. (2013). Development of a lab based epee fencing protocol. Int. J. Perform. Anal. Sport 13, 11–22. doi: 10.1080/24748668.2013.11868628

CrossRef Full Text | Google Scholar

Braeken, A. P., Kempen, G. I., Eekers, D., van Gils, F. C., Houben, R. M., and Lechner, L. (2011). The usefulness and feasibility of a screening instrument to identify psychosocial problems in patients receiving curative radiotherapy: a process evaluation. BMC Cancer 11:479. doi: 10.1186/1471-2407-11-479

PubMed Abstract | CrossRef Full Text | Google Scholar

Bridge, C. A., Ferreira da Silva Santos, J., Chaabene, H., Pieter, W., and Franchini, E. (2014). Physical and physiological profiles of taekwondo athletes. Sports Med. 44, 713–733. doi: 10.1007/s40279-014-0159-9

PubMed Abstract | CrossRef Full Text | Google Scholar

Castagna, C., Iellamo, F., Impellizzeri, F. M., and Manzi, V. (2014). Validity and reliability of the 45-15 test for aerobic fitness in young soccer players. Int. J. Sports Physiol. Perform. 9, 525–531. doi: 10.1123/ijspp.2012-0165

PubMed Abstract | CrossRef Full Text | Google Scholar

Chaabane, H., and Negra, Y. (2015). “Physical and physiological assessment,” in Karate Kumite: How to Optimize Performance, OMICS Group e-Book, ed H. Chaabane (Foster City, CA: Edt; OMICS Group incorporation).

Chaabène, H., Franchini, E., Sterkowicz, S., Tabben, M., Hachana, Y., and Chamari, K. (2015a). Physiological responses to karate specific activities. Sci. Sports 30, 179–187. doi: 10.1016/j.scispo.2015.03.002

CrossRef Full Text | Google Scholar

Chaabène, H., Hachana, Y., Attia, A., Mkaouer, B., Chaabouni, S., and Chamari, K. (2012a). Relative and absolute reliability of karate specific aerobic test (Ksat) in experienced male athletes. Biol. Sport 29. doi: 10.5604/20831862.1003485

CrossRef Full Text

Chaabène, H., Hachana, Y., Franchini, E., Mkaouer, B., and Chamari, K. (2012b). Physical and physiological profile of elite karate athletes. Sports Med. 42, 829–843. doi: 10.2165/11633050-000000000-00000

PubMed Abstract | CrossRef Full Text | Google Scholar

Chaabène, H., Hachana, Y., Franchini, E., Mkaouer, B., Montassar, M., and Chamari, K. (2012c). Reliability and construct validity of the karate-specific aerobic test. J. Strength Cond. Res. 26, 3454–3460. doi: 10.1519/JSC.0b013e31824eddda

PubMed Abstract | CrossRef Full Text | Google Scholar

Chaabène, H., Hachana, Y., Franchini, E., Tabben, M., Mkaouer, B., Negra, Y., et al. (2015b). Criterion Related Validity of Karate Specific Aerobic Test (KSAT). Asian J. Sports Med. 6:e23807. doi: 10.5812/asjsm.23807.

PubMed Abstract | CrossRef Full Text | Google Scholar

Chaabene, H., Negra, Y., Bouguezzi, R., Mkaouer, B., Franchini, E., Julio, U., et al. (2017a). Physical and physiological attributes of wrestlers: an update. J. Strength Cond. Res. 31, 1411–1442. doi: 10.1519/jsc.0000000000001738

PubMed Abstract | CrossRef Full Text | Google Scholar

Chaabene, H., Negra, Y., Capranica, L., Bouguezzi, R., Hachana, Y., Rouahi, M. A., et al. (2017b). Validity and reliability of a new test of planned agility in elite taekwondo athletes. J. Strength Cond. Res. doi: 10.1519/JSC.0000000000002325. [Epub ahead of print].

PubMed Abstract | CrossRef Full Text | Google Scholar

Chaabène, H., Tabben, M., Mkaouer, B., Franchini, E., Negra, Y., Hammami, M., et al. (2015c). Amateur boxing: physical and physiological attributes. Sports Med. 45, 337–352. doi: 10.1007/s40279-014-0274-7

PubMed Abstract | CrossRef Full Text | Google Scholar

Chen, C. Y., Dai, J., Chen, I. F., Chou, K. M., and Chang, C. K. (2015). Reliability and validity of a dual-task test for skill proficiency in roundhouse kicks in elite taekwondo athletes. Open Access J Sports Med 6, 181–189. doi: 10.2147/oajsm.s84671

PubMed Abstract | CrossRef Full Text | Google Scholar

Currell, K., and Jeukendrup, A. E. (2008). Validity, reliability and sensitivity of measures of sporting performance. Sports Med. 38, 297–316. doi: 10.2165/00007256-200838040-00003

PubMed Abstract | CrossRef Full Text | Google Scholar

da Silva Santos, J. F., and Franchini, E. (2016). Is frequency speed of kick test responsive to training? A study with taekwondo athletes. Sport Sci. Health 12, 377–382. doi: 10.1007/s11332-016-0300-2

CrossRef Full Text | Google Scholar

de Azevedo, P. H., Pithon-Curi, T., Zagatto, A. M., Oliveira, J., and Perez, S. (2014). Maximal lactate steady state in Judo. Muscles Ligaments Tendons J. 4, 132–136. doi: 10.11138/mltj/2014.4.2.132

PubMed Abstract | CrossRef Full Text | Google Scholar

Del Vecchio, F., Dimare, M., Franchini, E., and Schaun, G. (2014). Physical fitness and maximum number of all-out hikidashi uchi-komi in judo practitioners. Med. Dello Sport 67, 383–396.

Google Scholar

Edmondson, A. C., and McManus, S. E. (2007). Methodological fit in management field research. Acad. Manag. Rev. 32, 1246–1264. doi: 10.5465/AMR.2007.26586086

CrossRef Full Text | Google Scholar

Fliess-Douer, O., Vanlandewijck, Y. C., Lubel Manor, G., and Van Der Woude, L. H. (2010). A systematic review of wheelchair skills tests for manual wheelchair users with a spinal cord injury: towards a standardized outcome measure. Clin. Rehabil. 24, 867–886. doi: 10.1177/0269215510367981

PubMed Abstract | CrossRef Full Text | Google Scholar

Franchini, E., Del Vecchio, F. B., Matsushigue, K. A., and Artioli, G. G. (2011a). Physiological profiles of elite judo athletes. Sports Med. 41, 147–166. doi: 10.2165/11538580-000000000-00000

PubMed Abstract | CrossRef Full Text | Google Scholar

Franchini, E., Miarka, B., Matheus, L., and Vecchio, F. B. D. (2011b). Endurance in judogi grip strength tests: comparison between elite and non-elite judo players. Arch. Budo 7, 1–4.

Google Scholar

Franchini, E., Nakamura, F., Takito, M., Kiss, M., and Sterkowicz, S. (1998). Specific fitness test developed in Brazilian judoists. Biology of sport 15, 165–170.

Franchini, E., Takito, M., Kiss, M., and Strerkowicz, S. (2005). Physical fitness and anthropometrical differences between elite and non-elite judo players. Biol. Sport 22:315.

Google Scholar

Hachana, Y., Attia, A., Chaabène, H., Gallas, S., Sassi, R. H., and Dotan, R. (2012). Test-retest reliability and circadian performance variability of a 15-s Wingate Anaerobic Test. Biol. Rhythm Res. 43, 413–421. doi: 10.1080/09291016.2011.599634

CrossRef Full Text | Google Scholar

Haley, S. M., and Fragala-Pinkham, M. A. (2006). Interpreting change scores of tests and measures used in physical therapy. Phys. Ther. 86, 735–743.

PubMed Abstract | Google Scholar

Hopkins, W. G. (2000). Measures of reliability in sports medicine and science. Sports Med. 30, 1–15. doi: 10.2165/00007256-200030010-00001

PubMed Abstract | CrossRef Full Text | Google Scholar

Hopkins, W. G. (2002). Probabilities of clinical or practical significance. Sportscience 6, 589.

Hopkins, W. G., Marshall, S. W., Batterham, A. M., and Hanin, J. (2009). Progressive statistics for studies in sports medicine and exercise science. Med. Sci. Sports Exerc. 41, 3–13. doi: 10.1249/MSS.0b013e31818cb278.

PubMed Abstract | CrossRef Full Text | Google Scholar

Hulteen, R. M., Lander, N. J., Morgan, P. J., Barnett, L. M., Robertson, S. J., and Lubans, D. R. (2015). Validity and reliability of field-based measures for assessing movement skill competency in lifelong physical activities: a systematic review. Sports Med. 45, 1443–1454. doi: 10.1007/s40279-015-0357-0.

PubMed Abstract | CrossRef Full Text | Google Scholar

Impellizzeri, F. M., and Marcora, S. M. (2009). Test validation in sport physiology: lessons learned from clinimetrics. Int. J. Sports Physiol. Perform. 4, 269–277. doi: 10.1123/ijspp.4.2.269

PubMed Abstract | CrossRef Full Text | Google Scholar

International Olympic Committee (2017a). Available online at: https://www.olympic.org/the-ioc

International Olympic Committee (2017b). Statistics. (Accessed Nov. 6, 2017).

Khan, K. S., Kunz, R., Kleijnen, J., and Antes, G. (2003). Five steps to conducting a systematic review. J. R. Soc. Med. 96, 118–121. doi: 10.1177/014107680309600304

PubMed Abstract | CrossRef Full Text | Google Scholar

Kraemer, W. J., Clark, J. E., and Dunn-Lewis, C. (2012). “Athletes needs analysis. In NSCA's Guide to program design,“ in NSCA's Guide to Program Design, ed H. J. Champaign (Colorado: Human Kinetics).

Google Scholar

Lidor, R., Melnik, Y., Bilkevitz, A., Arnon, M., and Falk, B. (2005). Measurement of talent in judo using a unique, judo-specific ability test. J. Sports Med. Phys. Fitness 45, 32–37.

PubMed Abstract | Google Scholar

Miarka, B., Marques, J. B., and Franchini, E. (2011). Reinterpreting the history of women's judo in Japan. Int. J. Hist. Sport 28, 1016–1029. doi: 10.1080/09523367.2011.563633

PubMed Abstract | CrossRef Full Text | Google Scholar

Moher, D., Liberati, A., Tetzlaff, J., and Altman, D. G. (2009). Preferred reporting items for systematic reviews and meta-analyses: the PRISMA statement. Ann. Int. Med. 151, 264–269. doi: 10.7326/0003-4819-151-4-200908180-00135

PubMed Abstract | CrossRef Full Text

Morales, J., Franchini, E., Garcia-Masso, X., Solana-Tramunt, M., Busca, B., and Gonzalez, L. M. (2016). The work endurance recovery method for quantifying training loads in judo. Int. J. Sports Physiol. Perform. 11, 913–919. doi: 10.1123/ijspp.2015-0605

PubMed Abstract | CrossRef Full Text | Google Scholar

Morrow, J. Jr., Mood, D., Disch, J., and Kang, M. (2015). Measurement and Evaluation in Human Performance. Champaign, IL: Human Kinetics.

Google Scholar

Nunan, D. (2006). Development of a sports specific aerobic capacity test for karate - a pilot study. J. Sports Sci. Med. 5, 47–53.

PubMed Abstract | Google Scholar

Obminski, Z., Borkowski, L., and Sikorski, W. (2011). The shot put performance as a marker of explosive strength in polish amateur boxers. A pilot study. Arch. Budo 7, 173–177.

Oliveira, M. P., Szmuchrowski, L. A., Gomes Flor, C. A., Gonçalves, R., and Couto, B. P. (2015). “Correlation between the performance of taekwondo athletes in an Adapted Anaerobic Kick Test and Wingate Anaerobic Test,” in Proceedings of the 1st World Congress on Health and Martial Arts in Interdisciplinary Approach, HMA, ed R. M. Kalina (Warsaw: Archives of Budo), 130–134.

Google Scholar

Robertson, S. J., Burnett, A. F., and Cochrane, J. (2014). Tests examining skill outcomes in sport: a systematic review of measurement properties and feasibility. Sports Med. 44, 501–518. doi: 10.1007/s40279-013-0131-0.

PubMed Abstract | CrossRef Full Text | Google Scholar

Rocha, F., Louro, H., Matias, R., and Costa, A. (2016). Anaerobic fitness assessment in taekwondo athletes. A new perspective. Motricidade 12, 4–11.

Google Scholar

Roi, G. S., and Bianchedi, D. (2008). The science of fencing: implications for performance and injury prevention. Sports Med. 38, 465–481. doi: 10.2165/00007256-200838060-00003

PubMed Abstract | CrossRef Full Text | Google Scholar

Russell, M., Benton, D., and Kingsley, M. (2010). Reliability and construct validity of soccer skills tests that measure passing, shooting, and dribbling. J. Sports Sci. 28, 1399–1408. doi: 10.1080/02640414.2010.511247

PubMed Abstract | CrossRef Full Text | Google Scholar

Safrit, M. J., and Wood, T. M. (eds.). (1989). Measurement Concepts in Physical Education and Exercise Science. Champaign, IL: Human Kinetics Books. 45–72.

Google Scholar

Sant'Ana, J., Diefenthaeler, F., Dal Pupo, J., Detanico, D., Guglielmo, L. G. A., and Santos, S. G. (2014). Anaerobic evaluation of Taekwondo athletes. Int. Sport Med. J. 15, 492–499.

Sant'Ana, J., Franchini, E., Murias, J., and Diefenthaeler, F. (2017). Validity of a taekwondo specific test to measure vo2peak and the heart rate deflection point. J. Strength Cond. Res. doi: 10.1519/jsc.0000000000002153. [Epub ahead of print].

PubMed Abstract | CrossRef Full Text

Santos, L., Gonzalez, V., Iscar, M., Brime, J. I., Fernandez-Rio, J., Egocheaga, J., et al. (2010). A new individual and specific test to determine the aerobic-anaerobic transition zone (Santos Test) in competitive judokas. J. Strength Cond. Res. 24, 2419–2428. doi: 10.1519/JSC.0b013e3181e34774

PubMed Abstract | CrossRef Full Text | Google Scholar

Santos, L., Gonzalez, V., Iscar, M., Brime, J. I., Fernandez-Rio, J., Rodriguez, B., et al. (2011). Retesting the validity of a specific field test for judo training. J. Hum. Kinet. 29, 141–150. doi: 10.2478/v10078-011-0048-3

PubMed Abstract | CrossRef Full Text | Google Scholar

Santos, L., Gonzalez, V., Iscar, M., Brime, J. I., Fernandez-Rio, J., Rodriguez, B., et al. (2012). Physiological response of high-level female judokas measured through laboratory and field tests. Retesting the validity of the Santos test. J. Sports Med. Phys. Fitness 52, 237–244.

PubMed Abstract

Sertić, H., Vidranski, T., and Segedi, I. (2011). Construction and validation of measurement tools for the evaluation of specific agility in karate. Ido Mov. Culture. J. Martial Arts Anthrop. 11, 37–41.

Shiyan, V. V. (2011). A method for estimating special endurance in wrestlers. Int. J. Wrestl. Sci. 1, 24–32. doi: 10.1080/21615667.2011.10878916

CrossRef Full Text | Google Scholar

Smith, M. S., Dyson, R. J., Hale, T., and Janaway, L. (2000). Development of a boxing dynamometer and its punch force discrimination efficacy. J. Sports Sci. 18, 445–450. doi: 10.1080/02640410050074377

PubMed Abstract | CrossRef Full Text | Google Scholar

Sogabe, A., Maehara, K., Iwasaki, S., Sterkowicz-Przybycień, K., Sasaki, T., and Sterkowicz, S. (2015). “Correlation analysis between special judo fitness test and uchikomi shuttle run rest,” in Proceedings of the 1st World Congress on Health and Martial Arts in Interdisciplinary Approach, HMA, ed R. M. Kalina (Warsaw: Archives of Budo), 119–123.

Google Scholar

Sterkowicz, S., and Franchini, E. (2001). Specific fitness of elite and novice judoists. J. Hum. Kinetics 6, 81–98.

Google Scholar

Sterkowicz, S., and Franchini, E. (2009). Testing motor fitness in karate. Arch. Budo 5, 29–34.

Google Scholar

Stevens, B., and Gibbins, S. (2002). Clinical utility and clinical significance in the assessment and management of pain in vulnerable infants. Clin. Perinatol. 29, 459–468. doi: 10.1016/S0095-5108(02)00016-7

PubMed Abstract | CrossRef Full Text

Streiner, D. L. (1993). A checklist for evaluating the usefulness of rating scales. Can. J. Psychiatry 38, 140–148.

PubMed Abstract | Google Scholar

Streiner, D. L., and Norman, G. R. (2005). Health Measurement Scales: A Practical Guide to Their Development and Use, III Edn. Oxford: Oxford University Press.

Google Scholar

Streiner, D. L., Norman, G. R., and Cairney, J. (2015). Health Measurement Scales: A Practical Guide to Their Development and Use. New York, NY: Oxford University Press.

Google Scholar

Tabben, M., Coquart, J., Chaabene, H., Franchini, E., Chamari, K., and Tourny, C. (2014). Validity and reliability of new karate-specific aerobic test for karatekas. Int. J. Sports Physiol. Perform. 9, 953–958. doi: 10.1123/ijspp.2013-0465

PubMed Abstract | CrossRef Full Text | Google Scholar

Tavra, M., Franchini, E., and Krstulovic, S. (2016). Discriminant and factorial validity of judo-specific tests in female athletes. Arch. Budo 12, 93–99.

Thomas, J. R., Silverman, S., and Nelson, J. (2015). Research Methods in Physical Activity, 7E. Champaign, IL: Human Kinetics.

Google Scholar

Tsolakis, C., and Vagenas, G. (2010). Anthropometric, physiological and performance characteristics of elite and sub-elite fencers. J. Hum. Kinet. 23, 89–95. doi: 10.2478/v10078-010-0011-8

CrossRef Full Text | Google Scholar

Turner, A., Bishop, C., Chavda, S., Edwards, M., Brazier, J., and Kilduff, L. P. (2016). Physical characteristics underpinning lunging and change of direction speed in fencing. J. Strength Cond. Res. 30, 2235–2241. doi: 10.1519/jsc.0000000000001320

PubMed Abstract | CrossRef Full Text | Google Scholar

Utter, A., Goss, F., DaSilva, S., Kang, J., Suminski, R., Borsa, P., et al. (1997). Development of a wrestling-specific performance test. J. Strength Cond. Res. 11, 88–91.

Google Scholar

Weichenberger, M., Liu, Y., and Steinacker, J. M. (2012). A test for determining endurance capacity in fencers. Int. J. Sports Med. 33, 48–52. doi: 10.1055/s-0031-1284349

PubMed Abstract | CrossRef Full Text | Google Scholar

Weir, J. P. (2005). Quantifying test-retest reliability using the intraclass correlation coefficient and the SEM. J. Strength Cond. Res. 19, 231–240. doi: 10.1519/15184.1

PubMed Abstract | CrossRef Full Text | Google Scholar

Wright, G. A., Isaacson, M. I., Malecek, D. J., and Steffen, J. P. (2015). Development and assessment of reliability for a sandbag throw conditioning test for wrestlers. J. Strength Cond. Res. 29, 451–457. doi: 10.1519/jsc.0000000000000637

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: martial arts, validity, reliability, sensitivity, methodological quality, specific assessment

Citation: Chaabene H, Negra Y, Bouguezzi R, Capranica L, Franchini E, Prieske O, Hbacha H and Granacher U (2018) Tests for the Assessment of Sport-Specific Performance in Olympic Combat Sports: A Systematic Review With Practical Recommendations. Front. Physiol. 9:386. doi: 10.3389/fphys.2018.00386

Received: 09 January 2018; Accepted: 28 March 2018;
Published: 10 April 2018.

Edited by:

Luca Paolo Ardigò, University of Verona, Italy

Reviewed by:

Antonio Paoli, Università degli Studi di Padova, Italy
Giovanni Messina, University of Foggia, Italy

Copyright © 2018 Chaabene, Negra, Bouguezzi, Capranica, Franchini, Prieske, Hbacha and Granacher. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Helmi Chaabene, chaabanehelmi@hotmail.fr

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.