The number of stimuli required to reliably assess corticomotor excitability and primary motor cortical representations using transcranial magnetic stimulation (TMS): a systematic review and meta-analysis
Systematic Reviews volume 6, Article number: 48 (2017)
Transcranial magnetic stimulation (TMS) is a non-invasive means by which to assess the structure and function of the central nervous system. Current practices involve the administration of multiple stimuli over target areas of a participant’s scalp. Decreasing the number of stimuli delivered during TMS assessments would improve time efficiency and decrease participant demand. However, doing so may also compromise the within- or between-session reliability of the technique. The aim of this review was therefore to determine the minimum number of TMS stimuli required to reliably measure (i) corticomotor excitability of a target muscle at a single cranial site and (ii) topography of the primary motor cortical representation of a target muscle across multiple cranial sites.
Database searches were performed to identify diagnostic reliability studies published before May 2015. Two independent reviewers extracted data from studies employing single-pulse TMS to measure (i) the corticomotor excitability at a single cranial site or (ii) the topographic cortical organisation of a target muscle across a number of cranial sites. Outcome measures included motor evoked potential amplitude, map volume, number of active map sites and location of the map centre of gravity.
Only studies comparing the reliability of varying numbers of stimuli delivered to a single cranial site were identified. Five was the lowest number of stimuli that could be delivered to produce excellent within-session motor evoked potential (MEP) amplitude reliability (intraclass correlation coefficient (ICC) = 0.92, 95% CI 0.87 to 0.95). Ten stimuli were required to achieve consistent between-session MEP amplitudes among healthy participants (ICC = 0.89, 95% CI 0.76 to 0.95). However, between-session reliability was influenced by participant characteristics, intersession intervals and target musculature.
Further exploration of the reliability of multi-site TMS mapping is required. Five stimuli produce reliable MEP recordings during single-site TMS investigations involving one session. For single-site analyses involving multiple sessions, ten stimuli are recommended when investigating corticomotor excitability in healthy participants or the upper limb musculature. However, greater numbers of stimuli may be required for clinical populations or assessments involving the lower limb.
Systematic review registration
Transcranial magnetic stimulation (TMS) is a non-invasive means by which to assess the structure and function of the central nervous system. During TMS, a stimulation coil introduces a magnetic field over the target area of a participant’s scalp, inducing a secondary electrical current within cortical tissue . Using electromyography, the muscular activation resulting from motor cortex stimulation can be measured as motor evoked potentials (MEPs). The MEP data acquired during TMS provides insight into the excitability of a participant’s neuronal network and enables changes in cortex morphology to be tracked over time [1–3]. Over the last 15 years, TMS has evolved to include applications ranging from pre-surgical tumour mapping to investigating the neurophysiological effects of various rehabilitation approaches [2–5]. Current TMS practices involve the administration of multiple stimuli over target areas of a participant’s scalp [3, 6]. Stimuli are delivered as single pulses, usually with 3 to 5 s between each stimulus . Depending upon the desired neurophysiological index, stimuli may either be applied at a single cranial site, or systematically over a predefined grid, in a process known as ‘mapping’ .
Single-site analyses typically involve the administration of ten stimuli over the optimal cranial site, or ‘hotspot’ [6, 7]. During mapping, four to five stimuli are administered to each cranial site [3, 8]. Multiple stimuli are delivered in this manner in an attempt to minimise the influence of variations in experimenter’s technique and participant’s position . Further, increasing the number of stimuli delivered per site during TMS assessments may reduce the influence of inherent variations in an individual’s corticospinal activity [6–9]. However, delivering multiple stimuli in order to map the cortical representation of a particular muscle is time-consuming, restricting its use beyond the research environment . Further, prolonged TMS assessments have been associated with significant increases in participant’s fatigue and discomfort . This becomes particularly important when investigating neurological populations, where increased metabolic demands and poor baseline levels of endurance may limit adherence . Similarly, prolonged assessments are impractical when exploring conditions involving chronic pain . The importance of reducing data collection time is also evident from the observation that corticomotor excitability fluctuates with participants’ arousal and concentration, which may vacillate throughout prolonged investigations .
Reducing the number of stimuli delivered to each cranial site during TMS has the potential to improve the efficiency of the procedure while decreasing participants’ demand . However, lowering the number of stimuli per site may also compromise the procedure’s reliability . No reviews have systematically examined the reliability of outcomes obtained using varying numbers of stimuli per cranial site during TMS . Thus, the aims of this systematic review were to determine the number of TMS stimuli required to reliably measure (i) the corticomotor excitability of a target muscle at a single cranial site and (ii) the topography of the primary motor cortical representation of a target muscle across multiple cranial sites.
This systematic review was prepared in accordance with the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA)  (see Additional file 1) and A Measurement Tool to Assess Systematic Reviews (AMSTAR) . The methods were developed using items from the Cochrane Handbook for Systematic Reviews relevant to the reporting of systematic reviews of diagnostic reliability studies . The protocol for the systematic review was registered with the International Prospective Register of Systematic Reviews (PROSPERO) (registration number CRD42015024579) and has been published previously .
Searches were conducted in CINAHL, CENTRAL, EMBASE, MEDLINE, Neuroscience Information Framework (NIF), PEDro, PsycINFO, PubMed, Scopus and Web of Science databases from their inception to May 2015. Key words and medical subject headings (MeSH) related to ‘transcranial magnetic stimulation’, ‘cortical reorganisation’ and ‘reliability’ were used to identify relevant literature (see Additional file 2). Search strategies were developed in consultation with a librarian with systematic review expertise and customised to suit each database. Authors contacted experts in the field of TMS and searched in Google Scholar for additional studies. Searches of the International Clinical Trials Registry Platform (ICTRP) were also conducted to identify recently completed studies. The reference lists of all relevant articles were analysed to identify additional trials to be considered for inclusion.
Inclusion criteria were limited to full-text diagnostic reliability studies comparing the effect of the number of TMS stimuli on the within- or between-session reliability of either (i) corticomotor excitability at a single cranial site or (ii) the topographic cortical organisation of a target muscle across multiple cranial sites. Only studies employing single-pulse TMS were included. The number and position of targeted cranial coordinates were required to be consistent within and between sessions. Studies comparing the reliability of TMS outcomes obtained between cortical hemispheres, rather than within or between sessions using the same hemisphere, were excluded. Similarly, studies reporting only the ‘ideal’ number of stimuli at a particular site or presenting reliability predictions based upon mathematical models, rather than experimental testing, were excluded. Reliability studies that did not explore the effect of varying numbers of stimuli were also excluded. The outcome measures that were analysed included motor evoked potential amplitude, map volume, number of active map sites, location of the map centre of gravity (CoG) and distance between the centres of gravity of the target muscle and one or more neighbouring muscles.
Search results were exported to EndNote citation software (Version X4.0.2; Thomson Reuters EndNote, New York, NY, USA) for automatic duplicate removal. Any duplicates overlooked by the program were removed manually. Two independent reviewers then screened the exported articles for relevance by title and abstract. Potentially relevant papers were retrieved as full-text articles and assessed according to the eligibility criteria by these two reviewers. An additional reviewer was consulted to resolve uncertainty or disagreement regarding the eligibility of studies. Excluded studies and reasons for exclusion were then recorded. Further data extraction procedures can be found in the systematic review protocol .
Two tiers of methodological quality assessment were performed. First, the general experimental design of each study was appraised using a custom appraisal tool consisting of items from both the Quality Assessment of Reliability (QAREL) checklist  and the guidelines for assessing reliability studies developed by Bialocerkowski et al. . Second, the included studies were assessed according to the TMS-specific checklist developed by Chipchase et al. . The development and justification of these quality assessment tools is outlined in detail in the systematic review protocol .
Two reviewers independently assessed studies satisfying the eligibility criteria. For both general experimental design and TMS-specific methodology, items were scored as either present (1) or absent (0). In accordance with QAREL-based recommendations provided by Triano et al. , studies scoring four or less for general experimental design were deemed to be of low quality, studies scoring between five and seven were classified as moderate quality and studies scoring eight or above were classified as high quality. Any disagreements were resolved by a third reviewer.
Reliability estimates in the form of intraclass correlation coefficients (ICCs) were pooled using StatsDirect Software (version 3.0.177; StatsDirect Ltd, Cheshire, UK) . The ICC is the most accurate and commonly used indication of the size and direction of association between two variables . Intraclass correlation coefficients for within- and between-session reliability were interpreted using the following values: less than 0.50 = poor; 0.50 to 0.65 = moderate; 0.65 to 0.80 = good; and greater than 0.80 = excellent . Confidence intervals surrounding ICCs were calculated using the StatsDirect software, and significance was set at p < 0.05.
A standard correlation (Hedges-Olkin) random effects model, incorporating a Fisher’s r-to-z transformation, was used during meta-analyses. This random effects model utilised as methodological heterogeneity is inevitable during manual TMS assessments . The impact of heterogeneity was calculated using the I 2 statistic and interpreted as follows: 0 to 40% may be unimportant; 30 to 60% may represent moderate heterogeneity; 50 to 90% may represent substantial heterogeneity; and 75 to 100% represents considerable heterogeneity . In the cases of substantial methodological or statistical heterogeneity, subgroup analyses were performed in accordance with the study protocol.
Flow of studies
Initial database searches yielded 3712 potentially relevant papers (see Fig. 1). Four additional studies were retrieved from other sources. No ongoing or recently completed studies were retrieved via the International Clinical Trials Registry Platform (ICTRP). Following screening, four studies met the inclusion criteria (see Fig. 1).
Characteristics of included studies
Participants and assessment
Four studies contributed data on 78 participants, 51% of whom were male. The mean age of participants ranged from 27.5 to 76.0 years. Two studies investigated corticomotor excitability in the upper limb muscles [22, 23]: one study investigated the lower limb musculature  and one study explored the submental muscle group of the neck used during volitional swallowing  (see Table 1). As well as analysing data from healthy individuals, Lewis, Signal & Taylor  examined a clinical population in the form of individuals greater than 6 months post-cerebrovascular accident. This study was performed during active conditions, while all other included studies were performed under resting conditions. All included studies were based upon data from a single cranial site (see Table 1) and used manual (coordinate-based) coil positioning without neuronavigation.
General experimental design
The overall experimental design quality of the studies was moderate, with a mean quality appraisal score of 6/11. Only one study was deemed to be of high quality (score ≥ 8)  (see Table 2). This was the only study to state that all participants received the assessment as allocated and that all participants were included in statistical analyses. None of the included studies specified participant sampling methods or performed any form of assessor blinding (see Table 2).
TMS-specific protocol and reporting
All studies complied with the majority of items on the TMS-specific checklist designed by Chipchase et al. , with scores ranging from 16 to 23 out of a possible 26. However, as demonstrated in Table 3, three studies did not report on medication use, and two did not report on the presence of participant comorbidities (including neurological/psychiatric disorders). Further, no studies controlled the level of relaxation present in the muscles other than those being tested, and only one study reported on stimulation pulse shape, current direction, interstimulus intervals or participant arousal levels (see Table 3).
Meta-analyses and subgroup analyses
Four studies compared MEP amplitudes obtained at a single cranial coordinate using varying numbers of stimuli [22–25]. No multi-site (mapping) studies satisfying the eligibility criteria were identified.
Four studies compared the within-session reliability of MEP amplitudes recorded using varying numbers of stimuli administered to a single cranial site [22–25]. As demonstrated in Fig. 2, the pooled ICCs for within-session reliability gradually increased from 0.69 (95% CI 0.47 to 0.83) using three stimuli to 0.98 (95% CI 0.95 to 0.99) when using 15 stimuli. Five stimuli had significantly higher ICCs than three or four stimuli (p < 0.01). However, no significant differences were identified between five, ten and 15 stimuli (p = 0.29), which all demonstrated excellent within-session reliability (ICC > 0.80). However, the analyses exhibited considerable statistical heterogeneity (ranging from I 2 = 0% [95% CI 0 to 0%]) to I 2 = 60% [95% CI 0 to 84%]). The studies also had a large degree of methodological heterogeneity in terms of target musculature and clinical population. Subgroup analyses were therefore performed to account for the potential influence of these variables (see Table 4).
As shown in the subgroup analysis presented in Table 4, the trends observed in the overall meta-analysis remained when studies investigating healthy participants, elderly participants and the upper limb musculature were separated. This indicates that the results were not influenced by participants’ differences or methodological inconsistencies.
No statistically significant differences were observed with varying numbers of stimuli in studies investigating submental musculature or participants aged less than 65 years (see Table 4). However, in these studies, five still represented the lowest number of stimuli for which ICC confidence intervals were entirely within the ‘good to excellent’ reliability range. Intraclass correlation coefficients were also higher overall for these participant groups.
Four studies compared the between-session reliability of MEP amplitudes recorded using varying numbers of stimuli administered to a single cranial site [22–25]. While pooled ICCs for between-session reliability tended to increase as higher numbers of stimuli were administered (see Fig. 3), differences were not significant (p = 0.1). However, the analyses exhibited considerable statistical heterogeneity (ranging from I 2 = 0% [95% CI 0 to 0%] to I 2 = 69% [95% CI 0 to 86.1%]). The studies also had a large degree of methodological heterogeneity in terms of target musculature and time provided between successive sessions. Subgroup analyses were therefore performed to account for the potential influence of these variables (see Table 5).
As shown in Table 5, between-session reliability tended to increase with increasing numbers of stimuli, regardless of participant characteristics or target musculature. Overall, the ICCs for between-session reliability were significantly lower than those obtained for within-session reliability (p < 0.05). In healthy participants, ten stimuli were required to achieve excellent between-session reliability (ICC > 0.80). Studies involving submental musculature showed good reliability (ICC > 0.65) for all analyses, with no significant differences identified between varying numbers of stimuli. However, the results of these studies demonstrated wide confidence intervals, suggesting that higher numbers of stimuli may be required. Investigations of the upper limb musculature yielded good to excellent between-session reliability for analyses involving five, ten and 15 stimuli (see Table 5). Conversely, investigations of the lower limb musculature and individuals’ post-cerebrovascular accident (CVA) demonstrated poor between-session reliability (ICC < 0.65), even when ten stimuli were analysed. Intersession intervals less than 72 h yielded significantly higher ICCs than studies involving longer intersession intervals (p < 0.05) (see Table 5).
This review is the first to synthesise data from experimental studies comparing the within- and between-session reliability of TMS outcome measures obtained using varying numbers of stimuli per cranial site. Surprisingly, no eligible studies investigating multi-site mapping were identified. For single-site analyses, five was the lowest number of stimuli that could be delivered to produce excellent within-session MEP amplitude reliability. Increasing the number of stimuli beyond this value did not significantly increase reliability in any participant group. Conversely, between-session reliability was influenced by participant characteristics, target musculature and intersession intervals. The ICCs for between-session reliability were significantly lower than those obtained for within-session reliability, with a minimum of ten stimuli being required to achieve consistent between-session MEP amplitudes among healthy participants.
Although extensive database searches were employed, no studies investigating the reliability of varying numbers of stimuli per cranial site during multi-site TMS mapping were identified. However, single-site analyses, as explored in this review, have been used previously to inform mapping practices [26–28]. Most commonly, ten stimuli are administered over the ‘hotspot’ during single-site analyses . Although this approach has been validated using magnetic resonance imaging (MRI), the International Federation of Clinical Neurophysiology highlights that current recommendations were determined arbitrarily in an attempt to ensure high levels of reliability . The current review found no statistically significant differences between five and ten stimuli in terms of within-session MEP amplitude reliability, with both approaches exhibiting excellent intraclass correlation coefficients. This finding was consistent across all participant groups. Such results suggest that data acquisition times for single-site analyses involving one session may reliably be halved. Although single-site analyses are not as physiologically demanding as TMS mapping, reducing the number of stimuli required may enhance participants’ comfort. Importantly, it would also facilitate analysis of short-lived changes in corticomotor excitability, as well as exploration of the effects of interventions (such as sports taping or cortisone injections) on rapid neuroplasticity .
The between-session reliability of MEP amplitudes obtained following single-site TMS was also investigated. In healthy participants, ten stimuli were required to achieve excellent between-session reliability (ICC > 0.80). However, subgroup analyses revealed that this value was influenced by participant characteristics, target musculature, and intersession intervals. For example, greater numbers of stimuli were required to reliably assess participants’ post-cerebrovascular accident. Such findings are not surprising, as individuals with neurological lesions have been shown to exhibit significant variability in corticomotor excitability over time [28, 29]. Analyses of the lower limb musculature also required a greater number of stimuli to achieve high levels of between-session MEP amplitude reliability. This may be due to the difficulties associated with ensuring participants to maintain consistent submaximal contractions during TMS assessments involving the lower limb . As supported by mathematical models , greater intersession intervals were associated with lower ICCs. The number of stimuli required during TMS assessments therefore appears to be dependent upon the period of time between sessions.
While this systematic review and meta-analysis provides researchers with useful information regarding TMS investigations, the overall methodological quality of the evidence was classified as ‘moderate’ (general experimental design score of 6/11). All studies scored poorly in the areas of blinding and sampling methodology. Participant blinding is difficult during TMS investigations, and is often not a requirement during reliability studies in which subsets of scores are compared [15, 28]. However, blinding of assessors to participant characteristics and during data analysis is usually necessary [17, 18]. Random or consecutive sampling should also be utilised during reliability studies in order to reduce the potential for bias . As none of the studies employed these methodological processes, TMS procedures or sample characteristics may have been altered according to expected outcomes . Additionally, only one study explicitly stated that all participants were assessed and included in statistical analyses . Most of the included studies were therefore at risk of further bias associated with attrition or data exclusion [15–18].
All studies complied with the majority of items on the TMS-specific checklist  (scores ranging from 16 to 23 out of a possible 26). Thus, while general methodological designs may not have been robust, TMS-specific protocols were well implemented and reported. Despite this, three studies did not report on medication use [22–25], two did not report on the presence of participant comorbidities (including neurological/psychiatric disorders) [22, 23] and only one study reported on participant arousal . These factors have been shown to influence corticomotor excitability and participant adherence across TMS sessions [28, 29]. Further, no studies controlled the level of relaxation present in the muscles other than those being tested, and only one study reported on stimulation pulse shape, current direction, or interstimulus intervals. This makes it difficult to definitively determine if variations in MEP recordings over time are due to true corticomotor plasticity, or simply the result of inconsistent experimental procedures [29–31].
Despite a rigorous approach towards data collection and synthesis, this review is not without limitations. As database searches were limited to full-text articles, bias may have been introduced due to the exclusion of data in grey literature. This ‘publication bias’ may inflate reliability estimates, as studies with desirable or significant results are more likely to be granted publication . Additionally, including only one stimulation intensity per study (as per the review protocol) may have reduced the external validity of this review. The results of this study are therefore only applicable to the most commonly utilised stimulation intensities (110–130% of resting motor threshold). Another important limitation of the review is that it was not possible to investigate publication bias due to the limited number of articles that were retrieved for each comparison. While visual inspection of the forest plots suggests that the results were not greatly influenced by study sample size or publication date, the inability to produce funnel plots or perform formal tests of publication bias warrants further caution when interpreting the results. Similarly, this review highlighted a considerable amount of methodological and statistical heterogeneity across the included studies. Despite such limitations, the findings across the included studies were relatively consistent, even when subgroup analyses were performed. The majority of the studies were also recent, and TMS technology did not appear to vary significantly from study to study.
The findings of this review are also only applicable to ‘traditional’, systematic TMS approaches. Alternate mapping techniques, involving ‘pseudorandom’, rather than systematic, site stimulation, have shown promise in terms of improving the efficiency of TMS assessments . These approaches involve delivering single stimuli to each cranial site in a pseudorandom manner, reducing the need for repeated stimulation. However, the evidence supporting such techniques is still developing, and they have only been compared to abridged TMS protocols involving the delivery of three stimuli per cranial site, rather than traditional approaches involving five stimuli per cranial site . Further, the conventional systematic approach towards TMS assessment continues to be a common practice, so optimising the efficiency of this technique is an important pursuit.
While the number of studies exploring the efficiency of TMS is increasing, there remains a paucity of evidence on multi-site mapping. However, the results obtained from single-site analyses highlight the potential for reliably reducing the number of stimuli administered during TMS-based investigations. Future research should seek to determine if this potential can be translated to TMS mapping. Doing so would decrease participant demand during TMS investigations, while making data acquisition more achievable within a clinical setting and timeframe. Common indices of corticomotor excitability and organisation, including map volume, distance and number of active sites, should also be explored. The quality of future research may be improved by ensuring assessor blinding, and random sampling are performed. Studies should also utilise the TMS checklist  in order to provide sufficient methodological detail. This would ensure that variations in MEP recordings could more validly be attributed to changes in corticomotor excitability.
This review found that delivering five stimuli produces reliable MEP recordings during single-site TMS investigations involving one session. For single-site analyses involving multiple sessions, ten stimuli are recommended when investigating the corticomotor excitability in healthy participants or the upper limb musculature. However, further studies comparing varying numbers of stimuli during TMS assessments would be beneficial, as the methodological and statistical heterogeneity observed across the included studies warrants a degree of caution when interpreting the results. Greater numbers of stimuli may also be required for clinical populations or assessments involving the lower limb. If intersession intervals exceed 72 h, reliability may be compromised. Multi-site studies, including multiple indices of corticomotor excitability and exploration of clinical populations, are required to accurately determine the optimal number of stimuli for TMS mapping.
A Measurement Tool to Assess Systematic Reviews
Centre of gravity
Intraclass correlation coefficient
Motor evoked potential
Preferred Reporting Items for Systematic Reviews and Meta-Analyses
Quality Assessment of Reliability Studies
Transcranial magnetic stimulation
Barker AT, Jalinous R, Freeston IL. Non-invasive magnetic stimulation of human motor cortex. Lancet. 1985;325(8437):1106–7. doi:10.1016/s0140-6736(85)92413-4.
Reinacher P, Priebe H, Blumrich W, Zentner J, Scheufler K. The effects of stimulation pattern and sevoflurane concentration on intraoperative motor-evoked potentials. Anesth Analg. 2006;102(3):888–95. doi:10.1213/01.ane.0000195235.02162.5d.
Peterchev A, Wagner T, Miranda P, Nitsche M, Paulus W, Lisanby S, et al. Fundamentals of transcranial electric and magnetic stimulation dose: definition, selection, and reporting practices. Brain Stimul. 2012;5(4):435–53. doi:10.1016/j.brs.2011.10.001.
Rapisarda G, Bastings E, de Noordhout A, Pennisi G, Delwaide P. Can motor recovery in stroke patients be predicted by early transcranial magnetic stimulation? Stroke. 1996;27(12):2191–6. doi:10.1161/01.str.27.12.2191.
Stinear C. Prediction of recovery of motor function after stroke. Lancet Neurol. 2010;9(12):1228–32. doi:10.1016/s1474-4422(10)70247-7.
Groppa S, Oliviero A, Eisen A, Quartarone A, Cohen LG, Mall V, et al. A practical guide to diagnostic transcranial magnetic stimulation: report of an IFCN committee. Clin Neurophysiol. 2012;123(5):858–82. doi:10.1016/j.clinph.2012.01.010.
Schabrun SM, Hodges PW, Vicenzino B, Jones E, Chipchase LS. Novel adaptations in motor cortical maps: the relationship to persistent elbow pain. Med Sci Sports Exerc. 2014;5(1):1–34. doi:10.1249/mss.0000000000000469.
Mutanen T, Mäki H, Ilmoniemi R. The effect of stimulus parameters on TMS–EEG muscle artifacts. Brain Stimul. 2013;6(3):371–6. doi:10.1016/j.brs.2012.07.005.
Rossi S, Rossini P. TMS in cognitive plasticity and the potential for rehabilitation. Trends Cogn Sci. 2004;8(6):273–9. doi:10.1016/j.tics.2004.04.012.
van de Ruit M, Perenboom M, Grey M. TMS brain mapping in less than two minutes. Brain Stimul. 2015;8(2):231–9. doi:10.1016/j.brs.2014.10.020.
Mead G, Bernhardt J, Kwakkel G. Stroke: physical fitness, exercise, and fatigue. Stroke Res Treat. 2012;1(1):1–2. doi:10.1155/2012/632531.
Classen J, Knorr U, Werhahn K, Schlaug G, Kunesch E, Cohen L, et al. Multimodal output mapping of human central motor representation on different spatial scales. J Physiol. 1998;512(1):163–79. doi:10.1111/j.1469-7793.1998.163bf.x.
Moher D, Shamseer L, Clarke M, Ghersi D, Liberati A, Petticrew M, et al. Preferred reporting items for systematic review and meta-analysis protocols (PRISMA-P) 2015 statement. Syst Rev. 2015;4(1):1. doi:10.1186/2046-4053-4-1.
Jagannath V, Mathew J, Asokan G, Fedorowicz Z. Quality assessment of systematic reviews of health care interventions using AMSTAR. Indian Pediatr. 2010;48(5):383–5. doi:10.1007/s13312-011-0080-3.
Higgins J, Green S. Cochrane handbook for systematic reviews of interventions, version 5.0.2. Cochrane Collaboration. 2011. http://handbook.cochrane.org/. Accessed 20 Mar 2015.
Cavaleri R, Schabrun SM, Chipchase LS. Determining the number of stimuli required to reliably assess corticomotor excitability and primary motor cortical representations using transcranial magnetic stimulation (TMS): a protocol for a systematic review and meta-analysis. Syst Rev. 2015;4(107):1–5. doi:10.1186/s13643-015-0095-2.
Lucas N, Macaskill P, Irwig L, Bogduk N. The development of a quality appraisal tool for studies of diagnostic reliability (QAREL). J Clin Epidemiol. 2010;63(8):854–61. doi:10.1016/j.jclinepi.2009.10.002.
Bialocerkowski A, Klupp N, Bragge P. How to read and critically appraise a reliability article. Int J Ther Rehab. 2010;17(3):114–20. doi:10.12968/ijtr.2010.17.3.46743.
Chipchase L, Schabrun S, Cohen L, Hodges P, Ridding M, Rothwell J, et al. A checklist for assessing the methodological quality of studies using transcranial magnetic stimulation to study the motor system: an international consensus study. Clin Neurophysiol. 2012;123(9):1698–704. doi:10.1016/j.clinph.2012.05.003.
Triano J, Budgell B, Bagnulo A, Roffey B, Bergmann T, Copperstein R, et al. Review of methods used by chiropractors to determine the site for applying manipulation. Chiropr Man Therap. 2013;21(1):36. doi:10.1186/2045-709x-21-36.
StatsDirect Ltd. StatsDirect Software, Version 3.0.177, Chesire, UK. Available at http://www.statsdirect.com/
Bastani A, Jaberzadeh S. A higher number of TMS-elicited MEP from a combined hotspot improves intra- and inter-session reliability of the upper limb muscles in healthy individuals. PLoS One. 2012;7(10):1–8. doi:10.1371/journal.pone.0047582.
Christie A, Fling B, Mulwitz L, Kamen G. Reliability of motor-evoked potentials in the ADM muscle of older adults. Med Sci Sport Exer. 2006;38(1):522–40. doi:10.1249/00005768-200605001-02182.
Lewis G, Signal N, Taylor D. Reliability of lower limb motor evoked potentials in stroke and healthy populations: How many responses are needed? Clin Neurophysiol. 2014;125(4):748–54. doi:10.1016/j.clinph.2013.07.029.
Doeltgen S, Ridding M, O’Beirne G, Dalrymple-Alford J, Huckabee M. Test-retest reliability of motor evoked potentials (MEPs) at the submental muscle group during volitional swallowing. J Neurosci Meth. 2009;178:134–7. doi:10.1016/j.neumeth.2008.12.005.
Kamen G. Reliability of motor-evoked potentials during resting and active contraction conditions. Med Sci Sport Exer. 2004;36(9):1574–9. doi:10.1249/01.mss.0000139804.02576.6a.
Plowman-Prine E, Triggs W, Malcolm M, Rosenbek J. Reliability of transcranial magnetic stimulation for mapping swallowing musculature in the human motor cortex. Clin Neurophysiol. 2008;119(10):2298–303. doi:10.1016/j.clinph.2008.06.006.
Carroll T, Riek S, Carson R. Reliability of the input–output properties of the cortico-spinal pathway obtained from transcranial magnetic and electrical stimulation. J Neurosci Meth. 2001;112(2):193–202. doi:10.1016/s0165-0270(01)00468-x.
Martin P, Hudson A, Gandevia S, Taylor J. Reproducible measurement of human motoneuron excitability with magnetic stimulation of the corticospinal tract. J Neurophysiol. 2009;102(1):606–13. doi:10.1152/jn.91348.2008.
Lucas N, Macaskill P, Irwig L, et al. The reliability of a quality appraisal tool for studies of diagnostic reliability (QAREL). BMC Med Res Methodol. 2013;13(1):111. doi:10.1186/1471-2288-13-111.
Portney L, Watkins M. Foundations of clinical research: application to practice. 2nd ed. Melbourne: Pearson; 2009.
There are no additional acknowledgements to report.
This research received no specific grant from any funding agency in the public, commercial or not-for-profit sectors.
Availability of data and materials
The datasets analysed during the current study are available from the corresponding author upon reasonable request.
RC, LC and SMS were each involved in the conception, design, writing and editing of the study. The final review was approved by RC, LC and SMS. All authors read and approved the final manuscript.
The authors declare that they have no competing interests.
Consent for publication
Ethics approval and consent to participate
About this article
Cite this article
Cavaleri, R., Schabrun, S.M. & Chipchase, L.S. The number of stimuli required to reliably assess corticomotor excitability and primary motor cortical representations using transcranial magnetic stimulation (TMS): a systematic review and meta-analysis. Syst Rev 6, 48 (2017). https://doi.org/10.1186/s13643-017-0440-8