Scolaris Content Display Scolaris Content Display

Physical examination for lumbar radiculopathy due to disc herniation in patients with low‐back pain

Esta versión no es la más reciente

Referencias

Additional references

Bachmann 2002

Bachmann LM, Coray R, Estermann P, Ter Riet G. Identifying diagnostic studies in MEDLINE: reducing the number needed to read. Journal of the American Medical Informatics Association 2002;9(6):653‐8.

Bachmann 2003

Bachmann LM, Estermann P, Kronenberg C, ter Riet G. Identifying diagnostic accuracy studies in EMBASE. Journal of the Medical Library Association 2003;91(3):341‐6.

Boden 1990

Boden SD, Davis DO, Dina TS, Patronas NJ, Wiesel SW. Abnormal magnetic‐resonance scans of the lumbar spine in asymptomatic subjects. A prospective investigation. Journal of Bone and Joint Surgery 1990;72(3):403‐8.

Deeks 2001

Deeks JJ. Systematic reviews of evaluations of diagnostic and screening tests. In: Egger M, Davey Smith G, Altman DG editor(s). Systematic reviews in health care. London: BMJ Publishing Group, 2001:248‐82.

Deville 2000a

Devillé WL, van der Windt DA, Dzaferagic A, Bezemer PD, Bouter LM. The test of Lasegue: systematic review of the accuracy in diagnosing herniated discs. Spine 2000;25(9):1140‐7.

Deville 2000b

Devillé WL, Bezemer PD, Bouter LM. Publications on diagnostic test evaluation in family medicine journals: an optimal search strategy. Journal of Clinical Epidemiology 2000;53(1):65‐9.

Deville 2002

Devillé WL, Buntinx F, Bouter LM, Montori VM, de Vet HC, van der Windt DA, et al. Conducting systematic reviews of diagnostic studies: didactic guidelines. BMC Medical Research Methodology 2002;2:9.

Deyo 1990

Deyo RA, Loeser JD, Bigos SJ. Herniated lumbar intervertebral disk. Annals of Internal Medicine 1990;112(8):598‐603.

Deyo 1992

Deyo RA, Rainville J, Kent DL. What can the history and physical examination tell us about low back pain?. JAMA 1992;268(6):760‐5.

Glas 2003a

Glas AS, Roos D, Deutekom M, Zwinderman AH, Bossuyt PM, Kurth KH. Tumor markers in the diagnosis of primary bladder cancer. A systematic review. Journal of Urology 2003;169(6):1975‐82.

Glas 2003b

Glas AS, Lijmer JG, Prins MH, Bonsel GJ, Bossuyt PM. The diagnostic odds ratio: a single indicator of test performance. Journal of Clinical Epidemiology 2003;56(11):1129‐35.

Handbook 2005

Cochrane Handbook for Diagnostic Test Accuracy Reviews. July 2005.

Harbord 2007

Harbord RM, Deeks JJ, Egger M, Whiting P, Sterne JAC. A unification of models for meta‐analysis of diagnostic accuracy studiesy. Biostatistics 2007;8(2):239‐51.

Kim 2002

Kim YS, Chin DK, Yoon DH, Jin BH, Cho YE. Predictors of successful outcome for lumbar chemonucleolysis: analysis of 3000 cases during the past 14 years. Neurosurgery 2002;51(5 Suppl):S123‐8.

Kohlboeck 2004

Kohlboeck G, Greimel KV, Piotrowski WP, Leibetseder M, Krombholz‐Reindl M, Neuhofer R, et al. Prognosis of multifactorial outcome in lumbar discectomy: a prospective longitudinal study investigating patients with disc prolapse. Clinical Journal of Pain 2004;20(6):455‐61.

Littenberg 1993

Littenberg B, Moses LE. Estimating diagnostic accuracy from multiple conflicting reports: a new meta‐analytic method. Medical Decision Making 1993;13(4):313‐21.

Rebain 2002

Rebain R, Baxter GD, McDonough S. A systematic review of the passive straight leg raising test as a diagnostic aid for low back pain (1989 to 2000). Spine 2002;27(17):E388‐95.

Rebain 2003

Rebain R, Baxter GD, McDonough S. The passive straight leg raising test in the diagnosis and treatment of lumbar disc herniation: a survey of United kingdom osteopathic opinion and clinical practice. Spine 2003;28(15):1717‐24.

Reitsma 2005

Reitsma JB, Glas AS, Rutjes AW, Scholten RJ, Bossuyt PM, Zwinderman AH. Bivariate analysis of sensitivity and specificity produces informative summary measures in diagnostic reviews. Journal of Clinical Epidemiology 2005;58(10):982‐90.

Rutter 2001

Rutter CM, Gatsonis CA. A hierarchical regression approach to meta‐analysis of diagnostic test accuracy evaluations. Statistics in Medicine 2001;20(19):2865‐84.

van den Hoogen 1995

van den Hoogen HM, Koes BW, van Eijk JT, Bouter LM. On the accuracy of history, physical examination, and erythrocyte sedimentation rate in diagnosing low back pain in general practice. A criteria‐based review of the literature. Spine 1995;20(3):318‐27.

Vroomen 1999

Vroomen PC, de Krom MC, Knottnerus JA. Diagnostic value of history and physical examination in patients suspected of sciatica due to disc herniation: a systematic review. Journal of Neurology 1999;246(10):899‐906.

Whiting 2004

Whiting P, Rutjes AW, Dinnes J, Reitsma J, Bossuyt PM, Kleijnen J. Development and validation of methods for assessing the quality of diagnostic accuracy studies. Health Technology Assessment 2004;8(25):iii, 1‐234.

References to other published versions of this review

van der Windt 2008

van der Windt D, Simons E, Riphagen I, Ammendolia C, Verhagen A, Laslett M, Devillé W, Aertgeerts B, Deyo R, Bouter L, de Vet H. Physical examination for the diagnosis of lumbar radiculopathy due to disc herniation in patients with low‐back pain. Cochrane Database of Systematic Reviews 2007, Issue 4. [Chichester, UK: Wiley. Available from: http://www3.interscience.wiley.com/cgi‐bin/mrwhome/106568753/DTAP7.pdf]
Table 1. Items for the quality assessment of diagnostic accuracy studies (QUADAS)

Item and Guide to classification

1. Was the spectrum of patients representative of the patients who will receive the test in practice? Is it a selective sample of patients?  

Differences in demographic or clinical features between the study population and the source population may lead to selection bias or spectrum variation. In this item we will focus on selection bias: is a selective sample of patients included?

  • Classify as ‘yes’ if a consecutive series of patients or a random sample has been selected. Information should be given about setting, in‐ and exclusion criteria, and preferably number of patients eligible and excluded. If a mixed population of primary and secondary care patients is used: the number of participants from each setting is presented.

  • Classify as ‘no’ if healthy controls are used. Score also ‘no’ if non‐response is high and selective, or there is clear evidence of selective sampling. Score also ‘no’ if a population is selected that is otherwise unsuitable, for example, patients are known to have other specific causes of LBP (severe OA, malignancies, etc).

  • Classify as ‘unclear’ if insufficient information is given on the setting, selection criteria, or selection procedure to make a judgment. 

2. Is the reference standard likely to classify the target condition correctly?

Estimates of test performance are based on the assumption that the reference standard will identify nerve root compression due to disc herniation with 100% sensitivity and 100% specificity. Such reference standards are rare. Errors due to an imperfect reference standard may bias the estimation of diagnostic performance. For this review acceptable reference standards are: 1) findings at surgery demonstrating nerve root compression or irritation due to disc herniation; and 2) myelography indicating nerve root compression; and 3) although probably of lower quality, CT/MRI findings indicating nerve root compression;

  • Classify as ‘yes’ if one of these procedures is used as reference standards.

  • Classify as ‘no’ if you seriously question the methods used, if consensus among observers, or a combination of aspects of physical examination and history (‘clinical judgement’) is used as reference standard. (Use of imaging/surgery is actually a selection criterion, so the latter may not occur )

  • Classify as ‘unclear’ if insufficient information is given on the reference standard.

  • Classify as ‘not able’ if you consider yourself not capable to assess this item. If you have doubts, for example, regarding the quality of MRI‐procedures but feel not competent to make an adequate assessment, we can consult a radiologist.

3. Is the time period between the reference standard and the index test short enough to be reasonably sure that the target condition did not change between the two tests?

The index tests and reference standard should ideally be carried out at the same time. If there is a considerable delay, misclassification (due to spontaneous recovery or worsening of the condition) may occur.

  • Classify as ‘yes’ if the time period between physical examination and the reference standard is one week or less.

  • Classify as ‘no’ if the time period between physical examination and the reference standard is longer than one week.

  • Classify as ‘unclear’ if there is insufficient information on the time period between index tests and reference standard.  

4. Did the whole sample or a random selection of the sample receive verification using a reference standard of diagnosis?

When not all of the study patients receive confirmation of their diagnosis by a reference standard, partial verification bias may occur. Bias is very likely if the results of the index test influence the decision to perform the reference standard. Random allocation of patients to the reference standard should in theory not affect diagnostic performance. [Verification bias is also known as work‐up bias or sequential ordering bias].

  • Classify as ‘yes’ if it is clear that all patients who received the index test went on to receive a reference standard, even if the reference standard is not the same for all patients.

  • Classify as ‘no’ if not all patients who received the index test received verification by a reference standard.

  • Classify as ‘unclear’ if insufficient information is provided to assess this item.

5. Did patients receive the same reference standard regardless of the index test result?

Differential verification bias occurs when the results of the index tests are verified by different reference standards. This is not unlikely in this review: some patients may be referred for surgery following physical examination, whereas others only go on to receive diagnostic imaging. Bias is likely to occur when this decision depends on the results of the index test.

  • Classify as ‘yes’ if it is clear that all patients receiving the index test are subjected to the same reference standard.

  • Classify as ‘no’ if different reference standards are used.

  • Classify as ‘unclear’ if insufficient information is provided to assess this item.

6. Was the reference standard independent of the index test (i.e. the index test did not form part of the reference standard)?

It is not unlikely that the results of the physical examination are used when establishing the final diagnosis. In this case incorporation bias may occur (overestimating diagnostic accuracy). Knowledge of the results of the index test does not necessarily mean that these results are incorporated in the reference standard. For example, if the reference standard consists of MRI‐results only (regardless of knowledge of the results of the straight leg raising test), the index test is not part of the reference standard. However, if the final diagnosis is based on the results of both MRI‐findings and a positive straight leg raising test, incorporation bias will occur.

  • Score ‘yes’ if the index is no part of the reference standard.

  • Score ‘no’ if the index test is clearly part of the reference standard.

  • Score ‘unclear’ if insufficient information is provided to assess this item.

7. Were the index test results interpreted without knowledge of the results of the reference standard?

8. Were the reference standard results interpreted without knowledge of the results of the index test?

Interpretation of the results of physical examination may be influenced by knowledge of the results of the reference standard, and vice versa. This is known as reviewer bias, and may lead to over‐estimation of diagnostic accuracy. In our review the risk of bias may be substantial as both index test and reference standard often involve a subjective assessment of results. If the index test always precedes the reference standard, interpretation of the results of the index test will usually be without knowledge of the results of the reference standard. The reverse may also be true, although surgery is unlikely to precede physical examination.

  • Classify as ‘yes’ if the test results (index or reference standard) are interpreted blind to the results of the other test. Score also ‘yes’ if the sequence of testing is always the same and, consequently, one of the test is interpreted blind for the other.

  • Classify as ‘no’ if the assessor is aware of the results of the other test.

  • Classify as ‘unclear’ if insufficient information is given on independent or blind assessment of the index test or reference standard.

9. Were the same clinical data available when the index test results were interpreted as would be available when the test is used in practice?

The knowledge of demographic and clinical data, such as age, gender, symptoms, history of low‐backpain, previous treatments, or other aspects of physical examination may influence the interpretation of test results. The way this item is scored depends on the objective of the index test. If an aspect of physical examination is intended to replace other tests, these clinical data should not be available. However, if in practice clinical data are usually available when interpreting the results of the index test, this information should be available to the assessors of the index test.

  • Classify as ‘yes’ if clinical data (i.e. patient history, other physical tests) would normally be available when the test results are interpreted and similar data are available in the study.

  • Classify as ‘yes’ if clinical data would normally not be available when the test results are interpreted and these data are also not available in the study.

  • Classify as ‘no’ if this is not the case, e.g. if other test results are available that can not be regarded as part of routine care.

  • Classify as ‘unclear’ if the paper does not explain which clinical information was available at the time of assessment.

10. Were uninterpretable / intermediate test results reported?

Uninterpretable or intermediate test results are often not reported in diagnostic studies. Authors may simply remove these results from the analysis, which may lead to biased results of diagnostic performance. If uninterpretable or intermediate test results occur randomly and are not related to disease status, bias is unlikely. Whatever the cause of uninterpretable results they should be reported in order to estimate their potential influence on diagnostic performance.

  • Classify as ‘yes’ if all test results are reported for all patients, including uninterpretable, indeterminate or intermediate results.  

  • Classify as ‘yes’ if the authors do not report any uninterpretable, indeterminate or intermediate results AND the results are reported for all patients who were described as having been entered into the study.

  • Classify as ‘no’ if you think that such results occurred, but have not been reported.

  • Classify as ‘unclear’ if it is unclear whether all results have been reported.

11. Were withdrawals from the study explained?

Patients may withdraw from the study before the results of both index test and reference standard are known. If withdrawals systematically differ from patients remaining in the study, then estimates of diagnostic test performance may be biased. A flow chart is sometimes provided (in more recently published papers) which may help to score this item.

  • Classify as ‘yes’ if it is clear what happens to all patients who entered the study (all patients are accounted for, preferably in a flow chart).  

  • Classify as ‘yes’ if the authors do not report any withdrawals AND if the results are available for all patients who were reported to have been entered in the study.

  • Classify as ‘no’ if it is clear that not all patients who were entered completed the study (received both index test and reference standard), and not all patients are accounted for.

  • Classify as ‘unclear’ when the paper does not clearly describe whether or not all patients completed all tests, and are included in the analysis.

Note: In many diagnostic studies one may doubt whether or not all eligible patients have been entered in the study and are described in the paper. This issue is more strongly related to selection bias and will be scored under item 1.

Additional QUADAS items

12. Did the study provide a clear definition of what was considered to be a “positive” result of the index test?

Aspects of physical examination, for example the straight leg raising test, require a subjective judgement. Furthermore, several methods of performing the test have been described, and several cut‐offs have been proposed. Consequently, it is essential that an adequate description is given of the methods used to carry out (aspects of) physical examination, and how a positive result is defined.

  • Classify as ‘yes’ if the paper provides a clear description of the way the index test is performed, including a definition of a positive test result.

  • Classify as ‘no’ if no description is given of the way the index test is performed, and no definition is given of a positive test result.

  • Classify as ‘unclear’ if the methods of the index test are described, but no clear definition of a positive result has been provided, or vice versa.

13. Was treatment withheld until both index test and reference standard were performed?

If index tests and reference standard are not performed on the same day, some type of intervention may be initiated in between index test and reference standard. This might lead to misclassification (if some recovery of symptoms occurs).

  • Classify as ‘yes’ if no treatment is given in the time period between physical examination and the reference standard.

  • Classify as ‘no’ if an intervention is given that in your opinion could possibly influence the prognosis of low‐backpain due to nerve root compression / irritation.

  • Classify as ‘unclear’ if there is insufficient information regarding treatment between index test and reference standard.

14. Were data on observer variation reported and within acceptable range?

Studies on the reproducibility of physical examination in patients with musculoskeletal pain show that there may be considerable inter‐observer variation. This may strongly influence the diagnostic performance of the index test. It is difficult to give minimal cut‐off scores for inter‐observer agreement. A kappa or ICC of 0.70 is often considered to be acceptable, but this is certainly an arbitrary definition.

  • Classify as ‘yes’ if the paper provides information on inter‐observer variation, and the results are acceptable. 

  • Classify as ‘no’ if information is given on inter‐observer variation, and the results demonstrate poor agreement.

  • Classify as ‘unclear’ if there is insufficient information is provided regarding inter‐observer variation.

Each item is classified as “yes” (adequately addressed); “no” (inadequately addressed) or “unclear” (inadequate detail presented to allow a judgement to be made).

Figuras y tablas -
Table 1. Items for the quality assessment of diagnostic accuracy studies (QUADAS)