ArticlePDF Available

European guidelines on quality criteria for diagnostic radiographic images of the lumbar spine - An intra- and inter-observer reproducibility study

Authors:

Abstract and Figures

Background The Commission of the European Communities has published guidelines to be used as a gold standard for quality assessment of diagnostic radiographic images. Image quality and radiation dose must be monitored and optimally balanced for diagnostic purposes on patients. The objective of the current study was to assess intra- and inter-observer reproducibility in less experienced observers using the proposed European Guidelines on Quality Criteria for Diagnostic Radiographic Images in a quality assessment of lumbar spine radiographs in primary chiropractic practice in Denmark. Methods Two observers initially evaluated lumbar spine radiographs randomly selected from fifty chiropractic clinics, all connected to the national PACS server (KirPACS) in Denmark. All evaluations were performed twice by both observers using a four-week interval and for compliance with the European Quality Criteria for Diagnostic Radiographic Images. Inter- and intra-observer reproducibility was calculated using kappa statistics. In the interpretation of the kappa coefficient, the standards for strength of agreement reported by Landis and Koch were followed. Results The strength of the inter-observer agreement of general image quality at baseline ranged from moderate agreement (k = 0.47) to substantial agreement (k = 0.68). After four weeks, the inter-observer agreement still ranged from moderate agreement (k = 0.59) to substantial agreement (k = 0.71), but with increased agreement for both kappa coefficients. In relation to intra-observer agreement of general image quality, the strength for observer A ranged from moderate (k = 0.58) to substantial (k = 0.72) and the strength for observer B overall was substantial (k = 0.63–0.75). Conclusion The European Guidelines on Quality Criteria for Diagnostic Radiographic Images are considered a gold-standard and used in a method for quality assurance within the Danish chiropractic profession. The inter-rater and intra-rater agreements in this study, using the CEC-criteria, were found mostly acceptable. With appropriate attention to clear understanding of the individual criteria and sufficient training, this method is found to be reliable, even using less experienced observers, to carry out Diagnostic Radiographic Image Quality-assurance in primary care settings. Electronic supplementary material The online version of this article (10.1186/s12998-019-0241-3) contains supplementary material, which is available to authorized users.
Content may be subject to copyright.
R E S E A R C H Open Access
European guidelines on quality criteria for
diagnostic radiographic images of the
lumbar spine an intra- and inter-observer
reproducibility study
Klaus Doktor
1,2,4*
, Maria Lind Vilholm
2
, Aldis Hardardóttir
3
, Henrik Wulff Christensen
4
and Jens Lauritsen
5,6
Abstract
Background: The Commission of the European Communities has published guidelines to be used as a gold
standard for quality assessment of diagnostic radiographic images. Image quality and radiation dose must be
monitored and optimally balanced for diagnostic purposes on patients. The objective of the current study was to
assess intra- and inter-observer reproducibility in less experienced observers using the proposed European
Guidelines on Quality Criteria for Diagnostic Radiographic Images in a quality assessment of lumbar spine
radiographs in primary chiropractic practice in Denmark.
Methods: Two observers initially evaluated lumbar spine radiographs randomly selected from fifty chiropractic clinics,
all connected to the national PACS server (KirPACS) in Denmark. All evaluations were performed twice by both
observers using a four-week interval and for compliance with the European Quality Criteria for Diagnostic Radiographic
Images. Inter- and intra-observer reproducibility was calculated using kappa statistics. In the interpretation of the kappa
coefficient, the standards for strength of agreement reported by Landis and Koch were followed.
Results: The strength of the inter-observer agreement of general image quality at baseline ranged from moderate
agreement (k= 0.47) to substantial agreement (k= 0.68). After four weeks, the inter-observer agreement still ranged
from moderate agreement (k= 0.59) to substantial agreement (k= 0.71), but with increased agreement for both kappa
coefficients. In relation to intra-observer agreement of general image quality, the strength for observer A ranged from
moderate (k= 0.58) to substantial (k= 0.72) and the strength for observer B overall was substantial (k=0.630.75).
Conclusion: The European Guidelines on Quality Criteria for Diagnostic Radiographic Images are considered a gold-
standard and used in a method for quality assurance within the Danish chiropractic profession. The inter-rater and
intra-rater agreements in this study, using the CEC-criteria, were found mostly acceptable. With appropriate attention to
clear understanding of the individual criteria and sufficient training, this method is found to be reliable, even using less
experienced observers, to carry out Diagnostic Radiographic Image Quality-assurance in primary care settings.
Keywords: Agreement, Reliability, Reproducibility, EU-quality criteria, EU-guidelines, Lumbar spine, Radiographs, X-rays,
Radiography, Chiropractor, Imaging, Primary practice, Primary care
* Correspondence: k.doktor@nikkb.dk
1
Research Unit of Clinical Biomechanics, University of Southern Denmark,
Campusvej 55, 5250 Odense M, Denmark
2
Private chiropractic practice, Back Center Midwestern Jutland, Dalgas Allé
2A, 7400 Herning, Denmark
Full list of author information is available at the end of the article
© The Author(s). 2019 Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0
International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and
reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to
the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver
(http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.
Doktor et al. Chiropractic & Manual Therapies (2019) 27:20
https://doi.org/10.1186/s12998-019-0241-3
Background
In Denmark, primary chiropractic practices consist of
approximately 249 individual clinics [1], the vast majority
of which have their own radiographic imaging systems;
there are approximately 170 clinics using digitalized radio-
graphic systems, mostly Computed Radiography and to
less extent Direct Radiography [2].
Historically, chiropractors in Denmark have had the
rights and privileges to operate their own x-ray units. In
2008 a national Picture Archiving and Communication
System (PACS) was established at the Nordic Institute
for Chiropractic and Clinical Biomechanics (NIKKB),
University of Southern Denmark. The system (KirPACS)
was initially a standard PACS-system but has since been
developed and expanded with functionalities to include
documentation for various quality control and quality
assurance activities including a diagnostic second opin-
ion service for participating clinics. This unique system
is an example of a cost-efficient concept to monitor
radiation exposure doses and image quality control
procedures.
A protocol for quality assessment of lumbar spine
radiographs is proposed in the publication by the
Commission of the European Communities: EUR Report
16,260 European Guidelines on Quality Criteria for
Diagnostic Radiographic Images[3]. These guidelines
were used in a European-wide trial on the use of quality
criteria between the various professionals and authorities
involved in diagnostic radiology [4]. The image criteria
specify important anatomical structures that should be
visible in a radiograph to aide accurate diagnosis. Some
of these criteria depend fundamentally in correct posi-
tioning and cooperation of the patient, whereas others
reflect technical performance of the imaging system. A
qualitative guide to the necessary degree of visibility of
these essential structures is provided. They are a gold
standard for quality assessment of radiographic images
and must be used in any type of measures of image qual-
ity and dose relations (see Additional file 1: A and B). In
the original report the levels of agreement among ob-
servers are not clearly documented. This is also the case
for other similar studies [5].
Among Danish chiropractors, the European Guidelines
are used in an ongoing quality assurance program to be
performed once every 2 years. Studies of the reproduci-
bility of the image quality criteria are very limited in
numbers in the literature and we found it relevant to
report on our findings, since this is a crucial element in
optimizing the diagnostic gain for patients.
Objectives
The objective of this study was to assess the inter- and
intra-observer reproducibility in less experienced ob-
servers using an evaluation protocol conforming to the
European Guidelines on Quality Criteria for Diagnostic
Radiographic Imagesproposed by the European
Commission Study Group.
Methods
Design
The present study is an intra- and inter-observer repro-
ducibility study of the CEC-guidelines using repeated
measurements of individual lumbar spine radiographs.
Study population
To establish the level of intra- and inter-observer repro-
ducibility, data were extracted from readings of fifty
lumbar spine radiographs. The study materials were
randomly selected, anonymized and numbered from fifty
chiropractic clinics, retrieved from an archive of ap-
proximately 29,400 lumbar spine studies produced by
Danish chiropractors in 20152016 [2]. The focus was
purely on quality of images and was part of a quality
assurance program, as required biannually by Danish
law [6]. The study materials were blinded to the ob-
servers with respect to any personal information and no
diagnostic information was recorded.
Instruments
For the image analysis, the observers were using a digi-
tized format of the CEC Quality Criteria and all images
were retrieved from KirPACS using the image viewer
Osirix v. 5.7.1 for Mac [7]. The results were tabulated
directly into forms made in the software program
Epidata v.2.0.7.22 r547 [8].
Diagnostic monitors (2 million pixels) from Barco
(MDNC 2121 color led display) [9] were used for the
entire image evaluation process. Monitors had passed
acceptance tests according to Danish regulations.
Observers and training
The two observers were both licensed chiropractors by
the Danish National Board of Health and were in their
first 2 years of clinical practice. They were purposely
selected with limited clinical experience in accordance
with the study objective, but both had a high interest in
radiology in general. The observers received adequate
introduction in the use of the CEC quality criteria.
Initially ten lumbar spine series were evaluated in a joint
session to ensure consensus in understanding the
criteria and the evaluation process. An experienced
supervisor attended this initial session to guide consen-
sus. These studies were excluded from use in the repro-
ducibility study.
Blinding
The observers were blinded to any personal patient
information such as: Name, birth date, social security
Doktor et al. Chiropractic & Manual Therapies (2019) 27:20 Page 2 of 6
number, image accession number, report of findings and
clinic identification. They worked independently and
had no access to any previous readings or images. Every
effort was done to exclude any confounding factors that
could compromise our observers and the results. The
observers were given 2 weeks to finish their evaluations
and could log on and off to access the images any time
they wished.
General image quality assessment
After consensus was reached, fifty lumbar spine series
were evaluated by blinded and independent observers.
After more than four weeks, trying to minimize any rec-
ognition of image features, both observers re-evaluated
the same fifty image studies for the intra-observer
reliability evaluation. Observers could pick images in any
order, consecutive or random. The images were not
re-randomized for the second evaluation.
The general image quality assessment followed the scor-
ing principles described in the additional file 1.To
determine the level of reproducibility between the two ob-
servers, we used the General Image Acceptability-scores
for each of the three standard lumbar projections providing
three variables from each observer for each radiographic
series read and scored (See Additional file 1: A): 1.3.1
Acceptability Lumbar AP/PA-projection, 2.3.1. Acceptabil-
ity Lateral L1 to L4-projection and 3.3.1. Acceptability Lat-
eral L/S-projection. For description of the scoring
principles (see Additional file 1: B). General acceptability
was scored based on impressions of overall noise, contrast,
sharpness, collimation and patient positioning. Image
quality/acceptability were initially scored using a scale from
0 to 3 points (0 = unacceptable; 1 = only acceptable under
certain clinical conditions; 2 = probably acceptable, 3 = fully
acceptable).
It is important to establish an acceptable level of
agreement for the proposed method used to evaluate
diagnostic image quality. Any procedure used in evalua-
tions of performance must be validated to ensure reliable
results. We therefore tested the intra- and inter-observer
reproducibility of a general assessment of image quality
using Kappa-statistics.
Statistical analysis
For kappa statistics score-groups 0 and 1 were merged
into a not accepted groupand score-groups 2 or 3
were merged into an accepted group. All accepted
images received 1 point and not accepted images
received 0 point. This allowed us to calculate the intra-
and inter-observer reproducibility by means of ordinary
Kappa for binomial variables. Inter-observer reproduci-
bility was analyzed using results from the first (baseline)
evaluations. The ratings from each observer were
cross-tabulated in Epidata Entry Client and agreement
was measured using Cohens Kappa statistics in Stata.
Results were expressed as Kappa values with standard
errors and Z-scores indicated.
A Kappa value of 1 represents perfect agreement
between the observers; whereas a value of 0 means that
the results were obtained by chance. The Kappa values
were interpreted according to the recommendations of
Landis and Koch [10]. Values below 0.00 indicate poor
agreement; 0.000.20 slight agreement; 0.210.40 fair
agreement; 0.410.60 moderate agreement; 0.610.80
substantial agreement and a Kappa above 0.81 indicated
almost perfect agreement. Kappa values over 0.6 are
considered reliable.
Statistical analysis was performed using the STATA 14
for Windows, Stata Corporation, USA [11]; Microsoft Excel
2010, Microsoft Office Package, Microsoft Corporation,
USA [12]; Epidata Entry Client and Epidata Manager [8].
Results
A total of fifty lumbar spine radiographs were evaluated at
baseline by two independent observers. After 46 weeks,
the radiographs were re-evaluated by both observers to
determine the level of intra-observer reproducibility.
Inter-observer reproducibility
In Table 1, percent agreement, expected agreement and
Kappa values are presented at baseline.
In Table 2, percent agreement, expected agreement
and Kappa values are presented > 4 weeks.
Intra-observer reproducibility
In Table 3, percent agreement, expected agreement and
Kappa values for intra-observer agreements are pre-
sented for observers A and B.
General image quality/acceptability
The strength of the inter-observer agreement at baseline
ranged from moderate agreement (k= 0.47) to substan-
tial agreement (k= 0.68). After four weeks, the observers
read the images one more time and now the inter-
Table 1 Inter-observer agreement for general image quality assessment of the lumbar spine (baseline) n=50
Lumbar projection Agreement Expected Agreement Kappa (k) Standard Error Z Probability>Z
AP/PA 82.00% 62.48% 0.5203 0.1412 3.68 0.0001
Lat. L1-L4 84.00% 49.60% 0.6825 0.1374 4.97 0.0000
Lat. L5-S1 78.00% 58.28% 0.4727 0.1288 3.67 0.0001
Doktor et al. Chiropractic & Manual Therapies (2019) 27:20 Page 3 of 6
observer agreement increased although still ranging
from moderate agreement (k= 0.59) to substantial
agreement (k= 0.71).
In relation to intra-observer agreement of image
quality, the strength for observer A ranged from moder-
ate (k= 0.58) to substantial (k= 0.72) and the strength
for observer B overall was substantial (k= 0.630.75).
Kappa values > 0.6 is accepted as reliable.
Discussion
In this study of intra- and inter-observer reproducibility
when using the CEC-criteria to evaluate radiographic
image quality, we also wanted to find out if clinicians in
primary chiropractic practice can be expected to reach
acceptable levels of agreement, when used as observers.
Validating the use of the CEC-guidelines among clini-
cians with various degrees of experience is important,
since this method is a key element in the optimization of
image quality among chiropractors throughout Denmark
and because resources are allocated to continued
post-graduate education in image quality assurance.
Keeping in mind our objective, we intentionally left
observers with a brief introduction, instead of extensive
training in the use of image quality criteria in an attempt
to mimic a realistic clinical setting. We observed that
after the first set of fifty evaluations our observers
showed mildly improved levels of reliability. According
to the methods recommended by Landis and Koch, our
results can be rated with moderate to substantial
agreement.
Intra-observer reproducibility
Except for a borderline Kappa value for observer A for
the AP/PA lumbar projection, all other scores were
considered reliable. Since the two lateral projections are
traditionally combined into one image extending from
levels Th12-S2, in most chiropractic clinics, it is not
surprising that these two evaluations have almost identi-
cal Kappa scores. It is in full accordance with the
CEC-document that a standard lumbar spine series can
consist of only two, instead of three projections, al-
though this may require the use of a compression belt to
ensure even density.
The observers of this study were in their first years of
practice and had never worked with image quality
assessments, nevertheless it was possible to accomplish
mostly acceptable agreement. We would expect experi-
enced observers to achieve a higher level of agreement
with this evaluation system, as seen before in the publi-
cation by Maccia et al. [4].
Inter-observer reproducibility
If we exclude the first baseline Kappa values for inter-
observer reliability and use Kappa-values found at four
weeks, our results indicate that primary care clinicians
should be able to reliably apply this system with suffi-
cient initial training. A reason for higher levels of
disagreement in the first quality assessments was prob-
ably due to variations in perception of the image quality
criteria. The lateral lumbar spine projections achieved
better agreement than the AP/PA projection due to
better visibility of structures. The lumbar spine AP/PA
projection is more difficult to interpret, due to many
superimposed structures and the lumbar curve. Some
criteria are very clearly defined whereas others lack
sharp definitions leaving room for interpretation, e.g.,
important image details(Appendix B, 2.2) is sharply
defined for the lateral L1-L4 lumbar projection as:
Table 2 Inter-observer agreement for general image quality assessment of the lumbar spine (> 4 wks) n=50
Lumbar projection Agreement Expected Agreement Kappa (k) Standard Error Z Probability >Z
AP/PA 86.00% 66.00% 0.5882 0.1303 4.51 0.0000
Lat. L1-L4 86.00% 51.64% 0.7105 0.1370 5.19 0.0000
Lat. L5-S1 88.00% 60.76% 0.6942 0.1338 5.19 0.0000
Table 3 Intra-observer agreement for image quality evaluations of the lumbar spine
Agreement Expected Agreement Kappa (k) Standard Error Z Probability >Z
Observer A
AP/PA 86.00% 66.32% 0.5843 0.1350 4.33 0.0000
Lat. L1-L4 86.00% 49.52% 0.7227 0.1359 5.32 0.0000
Lat. L5-S1 90.00% 65.24% 0.7123 0.1352 5.27 0.0000
Observer B
AP/PA 86.00% 62.00% 0.6316 0.1358 4.65 0.0000
Lat. L1-L4 88.00% 51.20% 0.7541 0.1370 5.50 0.0000
Lat. L5-S1 84.00% 55.72% 0.6387 0.1356 4.71 0.0000
Doktor et al. Chiropractic & Manual Therapies (2019) 27:20 Page 4 of 6
Visually details down to 0.5 mm. at 3
rd
lumbar vertebral
body, ventral edge; whereas for the AP/PA lumbar
projection the definition is less specific: Visually details
down to 0.3-0.5 mm; and for the lateral L5/S1 lumbar
projection the definition is: Linear and reticular details
down to 0.5 mm. in width. It would likely strengthen
the Kappa scores, if we tightened the criteria interpreta-
tions by clarifying definitions (especially the AP/PA
projection) for future quality evaluations.
The repeated evaluations after four weeks increased
the agreement overall, which was concluded as likely to
be due to gained experience by the observers during the
study. It is important to allow observers to practise
the evaluations in an initial trial to reach consensus.
Our results indicate that more training of observers
prior to initiating the study could have improved
overall reliability.
In a study by Inah et al. of pelvic radiography image
quality in a Nigerian teaching hospital, image evaluations
were based on the Commission of European Communi-
ties (CEC) criteria, and an average Kappa value, k= 0.60
(0.360.76) for inter-observer reproducibility between
two radiologists was reported [13]. The study included
evaluation of 7 CEC quality criteria of the pelvis,
whereas in comparison, we used a general overall assess-
ment of lumbar spine image quality in our study and
found Kappa values ranging from, k= 0.470.68. In a
previous report of lumbar spine radiographic image
quality among Danish chiropractic clinics, we concluded
that a general quality assessment, as described above,
remained in good consistency with the results of evalua-
tions of 22 specific lumbar spine quality criteria. We
showed a correlation-coefficient, r = 0.720.83, in this
earlier study, indicating a clear positive correlation (un-
published report from NIKKB in 2000 by K. Doktor, N.
Grunnet-Nilsson and C. Lebouef-Yde). This further
emphasizes the usefulness of the criteria in a practical
clinic setting.
In another study of image quality Tesselaar et al. com-
pared lumbar spine radiographs in two different settings:
Sensitivity class 400 (less noise) and sensitivity class 560
(more noise). They concluded that higher image quality
produced higher inter-observer reliability, AC1 = 0.72 vs.
0.57 [14]. This is relevant to point out, since we, for
economic reasons, used an open source image viewer
for all image assessments.
There is limited data regarding the impact experience
has on the reliable assessment of image quality standards.
However, if we consider that experience in assessing
diagnostic images for quality offers some similarities to
assessing them for pathology (when using preset diagnos-
tic search criteria), then studies looking at the effect of
experience on reliability in diagnostic assessment may give
us insight into its effect on reliable image quality
assessment. Assendelft et al. found acceptable reproduci-
bility among Dutch chiropractors evaluating primarily
unspecific radiographic image findings. Intra-observer re-
liability was higher than inter-observer reliability [15].
Taylor et al. compared medical and chiropractic stu-
dents, clinicians and radiology specialists and found
higher reliability with more experienced observers
evaluating radiographs. Specialists obtained the best re-
sults with no differences between the two professions [16].
Similar results were found by de Zoete et al. [17]. This
suggests that greater experience in image assessment in
general is likely to increase reliability across a variety of
tasks.
Limitations
Our study used an open source viewer. It is possible that
if we had used a high-end image viewer our observers
would have obtained a higher inter-observer agreement.
Also, we didnt re-randomize image studies for the sec-
ond evaluation. This could possibly affect Kappa-scores
in a favorable way.
Conclusion and recommendations
The European Guidelines on Quality Criteria for Diag-
nostic Radiographic Images are considered a gold-stand-
ard and are used as a method for quality assurance
within the Danish chiropractic profession. The inter-ob-
server and intra-observer agreements in this study, using
the CEC-criteria, were found to be mostly acceptable.
With appropriate attention to a clear understanding of
the individual criteria and sufficient training this method
is found to be reliable, even using less experienced ob-
servers, to carry out Diagnostic Radiographic Image
Quality-assurance in primary care settings.
Our results indicate that primary care clinicians
should be able to reliably apply this system.
The CEC-quality criteria can be recommended for use
in any radiographic lumbar spine imaging setting. Our
results indicate that less experienced observers likely
would benefit from training in an initial trial of at least
50 imaging studies.
Additional file
Additional file 1: Overview of all lumbar image criteria, important
details, general assessment and examples. (PDF 652 kb)
Abbreviations
AP: From anterior to posterior; CEC: Council of the European Commission;
CR: Computed Radiography; DR: Direct Radiography; EU: European Union;
KirPACS: Danish Chiropractic Picture Archiving and Communication System;
L/S: Lumbo-sacral junction; L1: First lumbar vertebra; L4: Fourth lumbar
vertebra; L5: Fifth lumbar vertebra; NIKKB: Nordic Institute for Chiropractic
and Clinical Biomechanics; PA: From posterior to anterior; PACS: Picture
Archiving and Communication System; S1: First sacral vertebra
Doktor et al. Chiropractic & Manual Therapies (2019) 27:20 Page 5 of 6
Acknowledgements
The authors would like to acknowledge funding from the Foundation for
Chiropractic and Clinical Biomechanics. We also thank Orla Lund Nielsen for
providing us with kind assistance in handling data in Epidata. Last but not
least, we thank Sara Lisa Doktor for proofreading this paper.
Funding
Funding for the image analysis was provided from the Foundation for
Chiropractic and Clinical Biomechanics in Denmark.
Availability of data and materials
The datasets used and/or analyzed during the current study are available
from the corresponding author on reasonable request.
Authorscontributions
KD designed the study and performed the interpretation and analysis of
data and drafted the manuscript. AH and MLV contributed to the
conception of the study, evaluated the radiographs and entered results into
Epidata Entry Client. JL provided input to the definition and concept for
quality criteria and performed the statistical analysis. HWC supervised,
modified and proofread the manuscript. All authors read, critically reviewed
and approved the final version to be summited for publication.
Authorsinformation
Information on authors qualifications and affiliations is found on the first
page of this article.
Ethics approval and consent to participate
Not applicable. This study was done in conjunction with quality assurance
procedures required by Danish law [6] and all personal data were blinded for
the observers. Only documentation of image quality was performed, with no
possible correlation to patients. The procedures are mandatory and has no
consequences for diagnoses and treatment of patients.
Consent for publication
Not applicable.
Competing interests
The authors declare that they have no competing interests.
PublishersNote
Springer Nature remains neutral with regard to jurisdictional claims in
published maps and institutional affiliations.
Author details
1
Research Unit of Clinical Biomechanics, University of Southern Denmark,
Campusvej 55, 5250 Odense M, Denmark.
2
Private chiropractic practice, Back
Center Midwestern Jutland, Dalgas Allé 2A, 7400 Herning, Denmark.
3
Private
chiropractic practice, Reykjavik, Iceland.
4
Nordic Institute of Chiropractic and
Clinical Biomechanics, Campusvej 55, 5230 Odense M, Denmark.
5
Institute of
Clinical Medicine, University of Southern Denmark, Odense, Denmark.
6
Orthopedic Department, Odense University Hospital, Odense, Denmark.
Received: 8 October 2018 Accepted: 26 February 2019
References
1. Danish Chiropractor Association. KiroFAKTA 2016.
2. Nordic Institute of Chiropractic and Clinical Biomechanics. 2016.
3. Carmichael JHE, Maccia C, Moores BM, Oestmann JW, Schbilla H, et al.
European guidelines on quality criteria for diagnostic radiographic images:
EU publication EUR 16260; 2000.
4. Maccia C, Moores BM, Wall BF. The 1991 CEC trial on quality criteria for
diagnostic radiographic images: detailed results and findings; 1996.
5. Gonzalez L, Vano E, Oliete S, Manrique J, Hernáez JM, et al. Report of an
image quality and dose audit according to directive 97/43/Euratom at
Spanish private radiodiagnostics facilities. Br J Radiol. 1999;72:18692.
6. Sundhedsstyrelsen. Bekendtgørelse om medicinske røntgenanlæg til
undersøgelse af patienter, nr. 975 af 16. december 1998.
7. Synaptivemedical. Clear Canvas. http://www.clearcanvas.ca.
8. Lauritsen J, Bruus M. 2003-2018. Epidata. The Epidata Association, Odense,
Denmark. http://www.epidata.dk.
9. Barco. Nio 2MP (MDNC-2121) Specification sheet, Article number K9601651.
http://www.Barco.com. Belgium.
10. Landis JR, Koch GG. An application of hierarchical kappa-type statistics in
the assessment of majority agreement among multiple observers.
Biometrics. 1977;33:36374.
11. Statacorp. Stata statistical software: release 14. College Station, TX: StataCorp
LLC; 2015. http://www.stata.com
12. Microsoftcorp. Microsoft Office 2010 package. USA: Microsoft Corporation;
2018. http://www.microsoft.com.
13. Inah GB, Akintomide AO, Edim EE, Nzotta C, Egbe NO. A study of pelvic
radiography image quality in a Nigirian teaching hospital based on the
Commission of European Communities (CEC) criteria. South Afr Radiogr.
2013;51:159.
14. Tesselaar E, Dahlström N, Sandborg M. Clinical audit of image quality in
radiology using visual grading characteristics analysis. Radiat Prot Dosim.
2016;169:3406.
15. Assendelft WJ, Bouter LM, Knipschild PG, Wilmink JT. Reliability of lumbar
spine radiograph Reading by chiropractors. Spine. 1997;22(11):123541.
16. Taylor JAM, Clopton P, Bosch E, Miller K, Marcellis S. Interpretation of
abnormal Lumbosakral spine radiographs - a test comparing students,
clinicians, radiology residents and radiologists in medicine and chiropractic.
Spine. 1995;20:114753.
17. de Zoete A, Assendelft WJ, Algra PR, Oberman WR, Vanderschueren GM,
Bezemer PD. Reliability and validity of lumbosacral spine radiograph reading
by chiropractors, chiropractic radiologists, and medical radiologists. Spine.
2002;27(17):192633.
Doktor et al. Chiropractic & Manual Therapies (2019) 27:20 Page 6 of 6
... A previous report of a pilot study concluded that this method was suitable and recommended for radiographic image quality assurance programs within the Danish chiropractic profession [7]. This paper is following a recent publication on the reproducibility of the use of the CEC Quality Criteria [8]. ...
... The inter-and intra-observer reproducibility study included 50 radiographic studies of the lumbar spine and has been reported in full detail in a separate paper [8]. ...
Article
Full-text available
Background The Commission of the European Communities (CEC) has published: European Guidelines on Quality Criteria for Diagnostic Radiographic Images. These guidelines are considered a gold standard, recommended for use in quality assurance protocols. The objectives of this study: 1) Propose a graded classification format for Danish chiropractic clinics according to the CEC-quality criteria for diagnostic radiographic images of the lumbar spine. 2) Propose a reporting principle for quality of radiographic images. 3) Document variation in radiation exposure among clinics. Methods This is a cross-sectional study of image quality based on random sampling from 148 chiropractic clinics. Clinics were included if using: 1) Digital radiography and 2) The chiropractic picture and archiving system (KirPACS) at the Nordic Institute of Chiropractic and Clinical Biomechanics (NIKKB) in Denmark. A sample of 296 lumbar spine series were randomly collected from KirPACS (January 2018). Two independent observers reviewed 50 lumbar spine series twice with a 4-week interval, testing intra- and inter-observer reproducibility. The same observers then reviewed the remaining 246 radiographic studies. All studies were evaluated using the CEC Quality Criteria. Patient radiation dose values were retrieved from KirPACS (First quarter of 2020). Results A reporting and classification principle of diagnostic image quality was used in 148 chiropractic clinics. Compliance with the 22 CEC Quality Criteria had proportions ranging from 0.72–0.96 for 18 criteria, while 4 criteria specifying detail and definition ranged between 0.20–0.66. The proposed rating system (A to E) revealed: 18 A clinics, 28 B clinics, 32 C clinics, 25 D clinics and 45 E clinics (A = highest quality; E = lowest quality). The patient radiation reference dose in Denmark is 7 mGy for the AP/PA lumbar spine. Very few clinics exceed the reference dose value, approximately 50% of clinics were below 5 mGy. Conclusion A reporting principle is proposed for a graded classification format based on the CEC-quality criteria for diagnostic radiographic images of the lumbar spine. The Quality Criteria are for the most part met satisfactorily in 148 Danish chiropractic clinics, but important image details are compromised, in most cases, because of low patient radiation doses. The results of a patient radiation dose survey enabled documentation of variation in radiation exposure among chiropractic clinics.
... A previous report of a pilot study concluded that this method was suitable and recommended for radiographic image quality assurance programs within the Danish chiropractic profession (11). This paper is following a recent publication on the reproducibility of the use of the CEC Quality Criteria (12). ...
... The inter-and intra-observer reproducibility study included 50 radiographic studies of the lumbar spine and have been reported in full detail in a separate paper (12). ...
Preprint
Full-text available
Background The Commission of the European Communities (CEC) has published: European Guidelines on Quality Criteria for Diagnostic Radiographic Images. These guidelines are considered a gold standard, recommended for use in quality assurance protocols. The objectives of this study: 1) Propose a graded classification format for Danish chiropractic clinics according to the CEC-quality criteria for diagnostic radiographic images of the lumbar spine. 2) Propose a reporting principle for quality of radiographic images. 3) Document variation in radiation exposure among clinics.Methods This is a cross-sectional study of image quality based on random sampling from 148 chiropractic clinics. Clinics were included if using: 1) Digital radiography and 2) The chiropractic picture and achieving system(KirPACS) at the Nordic Institute of Chiropractic and Clinical Biomechanics(NIKKB) in Denmark. A sample of 296 lumbar spine series were randomly collected from KirPACS. Two independent observers reviewed 50 lumbar spine series twice with a 4-week interval, testing intra- and inter-observer reproducibility. The same observers then reviewed the remaining 246 radiographic studies. All studies were evaluated up against the CEC Quality Criteria. Patient radiation dose values were retrieved from KirPACS. ResultsA reporting and classification principle of diagnostic image quality was used in 148 chiropractic clinics. Compliance with the 22 CEC Quality Criteria had proportions ranging from 0.72-0.96 for 18 criteria, while 4 criteria specifying detail and definition ranged between 0.20-0.66. The proposed rating system (A to E) revealed: 18 A clinics, 28 B clinics, 32 C clinics, 25 D clinics and 45 E clinics (A = highest quality; E= lowest quality). The patient radiation reference dose in Denmark is 7 mGy for the AP/PA lumbar spine. Very few clinics exceed the reference dose value, approximately 50 % of clinics were below 5 mGy.ConclusionA reporting principle is proposed for a graded classification format based on the CEC-quality criteria for diagnostic radiographic images of the lumbar spine. The Quality Criteria are for the most part met satisfactory in 148 Danish chiropractic clinics, but important image details are compromised, in most cases, because of low patient radiation doses. The results of a patient radiation dose survey enabled documentation of variation in radiation exposure among chiropractic clinics.
... The first round of scoring was based on the detectability of clinically relevant anatomical structures. These anatomical structures were adopted and modified from the European Guidelines on Quality Criteria for Diagnostic Radiographic images, 27 omitting the positional criteria. The criteria were further modified following discussions with a group of experts. ...
Article
Full-text available
Background Radiographic X‐ray imaging is a common clinical examination. Current objective methods for quantifying image quality for radiographs struggle to capture the combined impact of factors throughout the imaging chain on the perceived image quality. Therefore, there is a need to further develop metrics that correlate with image quality as perceived by the observer. Objectives We proposed the image feature index (IFI) to comprehensively quantify radiographic X‐ray image quality. We also aimed to study the correlation between IFI and observer‐perceived image quality for chest radiographs. Materials and methods The IFI algorithm was developed, which measured the amount of information, textural features, and noise in the image. A total of 70 chest phantom radiographs were generated under 60–120 kV and 0.2–80 mAs. A vendor‐proprietary exposure index (EI) and dose area product (DAP) were extracted from the DICOM header in addition to calculating IFI for each image to investigate the relationships between IFI, EI and DAP. The quality of the images was rated by three observers, and the correlation between IFI and subjective score of image quality was tested. Next, a retrospective study using a random sample of 50 clinical chest radiographs was performed, and the correlation between IFI and subjective score was tested. The correlation was determined by the Spearman test. Results The curves of IFI versus DAP and IFI versus EI both demonstrated a similar three‐stage form where IFI is above zero: in the first stage, IFI increases rapidly with increased DAP or EI, whereas in the second stage, the slope of the curves decreased towards an asymptote, that is, minimal gain in IFI with increased DAP or EI—until they hit the inflection point and then descended sharply in the third stage. For both phantom and clinical chest images, IFI demonstrated good correlation with subjective score (r = 0.9084 for phantom images, r = 0.8153 for clinical images). Conclusions IFI is a feasible and efficient descriptor for image quality for chest radiographs. Future studies with larger sample sizes and sample types are needed to confirm the feasibility of IFI for other exam types and anatomical views, thus fulfilling and extending the potential applications of IFI in quality control and radiation dose reduction.
... The interpretation of all images were performed by three raters, selected from the department: a medical radiologist consultant with 30 years of experience in musculoskeletal MRI (rater A); a Ph.D. student with 28 years of clinical and radiography experience and 4 years of MRI experience including 1000 supervised spinal MRI reports (rater B), and a senior researcher with 12 years of clinical research and MRI experience, including 1000 supervised spinal MRI reports (rater C). All raters had experience with reliability studies and diagnostic classification models in diagnostic imaging [10,[13][14][15][16]. ...
Article
Full-text available
Objective To determine the inter-rater reliability of identifying differences and types of differences in lumbar degenerative findings comparing supine and upright MRI. Materials and methods Fifty-nine participants, low back pain patients (LBP) with or without leg pain and no-LBP individuals were consecutively enrolled to receive supine and upright MRI of the lumbar spine. Three raters independently evaluated the MRIs for degenerative spinal pathologies and compared for differences. Presence/absence of degenerative findings were recorded for all supine and upright images, and then differences from the supine to the upright positions were classified into no-change, appeared, disappeared, worsened, or improved at each individual disc level. Reliability and agreement were calculated using Gwet’s agreement coefficients (AC 1 or AC 2 ) and absolute agreement. Results Inter-rater reliability of evaluating differences in eight degenerative lumbar findings comparing the supine and upright MRI position, ranged from 0.929 to 0.996 according to Gwet’s agreement coefficients (AC 2 ). The total number of positive MRI findings in the supine position ranged from 270 to 453, with an average of 366 per rater. Observed differences from supine to upright MRI ranged from 18 to 80, with an average of 56 per rater. Conclusion Inter-rater reliability was found overall acceptable for classification of differences in eight types of degenerative pathology observed with supine and upright MRI of the lumbar spine. Results were primarily driven by high numbers and high reliability of rating negative findings, whereas agreement regarding positive findings and positive positional differences was lower.
... Image quality was graded using a scoring chart generated using image quality scoring criteria and other guidelines. [17][18][19] The scores were as follows: 1 = excellent, 2 = more than adequate, 3 = adequate, 4 = less than adequate, 5 = unacceptable, and 6 = incomplete. The radiologists scored each combination of initial and secondary study interpretation based on the American College of Radiology's RADPEER scoring system and categorized the interpretation discrepancies in each case to identify the discrepancy source. ...
Article
Purpose The purpose of this retrospective study was to evaluate the quality of outside hospital imaging and associated reports submitted to us for reinterpretation related to clinical care at our tertiary cancer center. We compared the initial study interpretations to that of interpretations performed by subspecialty-trained abdominal radiologists at our center and whether this resulted in a change in inpatient treatment. Materials and methods We performed an institutional review board-approved retrospective single-institution study of 915 consecutive outside computed tomography (CT) and magnetic resonance (MR) abdominal imaging studies that had been submitted to our institution between August 1, 2020 and November 30, 2020. The assessed parameters included the quality and accuracy of the report, the technical quality of the imaging compared to that at our institution, the appropriateness of the imaging for staging or restaging, usage of oral and IV contrast, and CT slice thickness. Clinical notes, pathologic findings, and subsequent imaging were used to establish an accurate diagnosis and determine the effect on clinical treatment. Discrepancies between the initial and secondary interpretations were identified independently by a panel of radiologists to assess changes in treatment. The impact of discrepancies on treatment was evaluated based on current treatment guidelines. Results Of 744 CT (81%) and 171 MR (19%) outside imaging studies, 65% had suboptimal quality compared to the images at our institution, and 31% were inappropriate for oncological care purposes. Only 21% of CT studies had optimal slice thickness of <3 mm. Of 375 (41%) outside reports, 131 (34%) had discrepancies between secondary and initial interpretations. Of the 88 confirmed discrepant studies, 42 patients (48%) had a change in treatment based on the secondary interpretation. Conclusions Imaging studies from outside institutions have variable image quality and are often inadequate for oncologic imaging. The secondary interpretations by subspecialty-trained radiologists resulted in treatment change.
... Rater 1, a medical radiologist consultant with 30 years of experience in musculoskeletal MRI; Rater 2, a chiropractor and PhD student with 28 years of clinical and radiography experience and 4 years of MRI experience including over 1000 supervised reports of lumbar MRI in the same radiology department; Rater 3, a chiropractor and senior researcher with 12 years of clinical research and MRI experience from radiology departments. All raters had, in various degrees, experience with reliability of diagnostic classification model as well as clinical experience with spinal diagnostic imaging [13,17,20,21]. ...
Article
Full-text available
Background: For diagnostic procedures to be clinically useful, they must be reliable. The interpretation of lumbar spine MRI scans is subject to variability and there is a lack of studies where reliability of multiple degenerative pathologies are rated simultaneously. The objective of our study was to determine the inter-rater reliability of three independent raters evaluating degenerative pathologies seen with lumbar spine MRI. Methods: Fifty-nine people, 35 patients with low back pain (LBP) or LBP and leg pain and 24 people without LBP or leg pain, received an MRI of the lumbar spine. Three raters (one radiologist and two chiropractors) evaluated the MRIs for the presence and severity of eight degenerative spinal pathologies using a standardized format: Spondylolisthesis, scoliosis, annular fissure, disc degeneration, disc contour, nerve root compromise, spinal stenosis and facet joint degeneration. Findings were identified and classified at disc level according to type and severity. Raters were instructed to evaluate all study sample persons once to assess inter-rater reliability (fully crossed design). Reliability was calculated using Gwet's Agreement Coefficients (AC1 and AC2) and Cohen's Kappa (κ) and Conger's extension of Cohen's. Gwet's probabilistic benchmarking method to the Landis and Koch scale was used. MRI-findings achieving substantial reliability was considered acceptable. Results: Inter-rater reliability for all raters combined, ranged from (Gwet's AC1 or AC2): 0.64-0.99 and according to probabilistic benchmarking to the Landis and Koch scale equivalent to moderate to almost perfect reliability. Overall reliability level for individual pathologies was almost perfect reliability for spondylolisthesis, spinal stenosis, scoliosis and annular fissure, substantial for nerve root compromise and disc degeneration, and moderate for facet joint degeneration and disc contour. Conclusion: Inter-rater reliability for 3 raters, evaluating 177 disc levels, was found to be overall acceptable for 6 out of 8 degenerative MRI-findings in the lumbar spine. Ratings of facet joint degeneration and disc contour achieved moderate reliability and was considered unacceptable.
Article
Full-text available
The aim of this work was to assess whether an audit of clinical image quality could be efficiently implemented within a limited time frame using visual grading characteristics (VGC) analysis. Lumbar spine radiography, bedside chest radiography and abdominal CT were selected. For each examination, images were acquired or reconstructed in two ways. Twenty images per examination were assessed by 40 radiology residents using visual grading of image criteria. The results were analysed using VGC. Inter-observer reliability was assessed. The results of the visual grading analysis were consistent with expected outcomes. The inter-observer reliability was moderate to good and correlated with perceived image quality (r2 = 0.47). The median observation time per image or image series was within 2 min. These results suggest that the use of visual grading of image criteria to assess the quality of radiographs provides a rapid method for performing an image quality audit in a clinical environment.
Article
Full-text available
A cross-sectional diagnostic study was conducted in two sessions. To determine and compare the reliability and validity of contraindications to chiropractic treatment (infections, malignancies, inflammatory spondylitis, and spondylolysis-listhesis) detected by chiropractors, chiropractic radiologists, and medical radiologists on plain lumbosacral radiographs. Plain radiography of the spine is an established part of chiropractic practice. Few studies have assessed the ability of chiropractors to read plain radiographs. Five chiropractors, three chiropractic radiologists and five medical radiologists read a set of 300 blinded lumbosacral radiographs, 50 of which showed an abnormality (prevalence, 16.7%), in two sessions. The results were expressed in terms of reliability (percentage and kappa) and validity (sensitivity and specificity). The interobserver agreement in the first session showed generalized kappas of 0.44 for the chiropractors, 0.55 for the chiropractic radiologists, and 0.60 for the medical radiologists. The intraobserver agreement showed mean kappas of 0.58, 0.68, and 0.72, respectively. The difference between the chiropractic radiologists and medical radiologists was not significant. However, there was a difference between the chiropractors and the other professional groups. The mean sensitivity and specificity of the first round, respectively was 0.86 and 0.88 for the chiropractors, 0.90 and 0.84 for the chiropractic radiologists, and 0.84 and 0.92 for the medical radiologists. No differences in the sensitivities were found between the professional groups. The medical radiologists were more specific than the others. Small differences with little clinical relevance were found. All the professional groups could adequately detect contraindications to chiropractic treatment on radiographs. For this indication, there is no reason to restrict interpretation of radiographs to medical radiologists. Good professional relationships between the professions are recommended to facilitate interprofessional consultation in case of doubt by the chiropractors.
Article
This paper presents a general statistical methodology for the analysis of multivariate categorical data involving agreement among more than two observers. Since these situations give rise to very large contingency tables in which most of the observed cell frequencies are zero, procedures based on indicator variables of the raw data for individual subjects are used to generate first-order margins and main diagonal sums from the conceptual multidimensional contingency table. From these quantities, estimates are generated to reflect the strength of an internal majority decision on each subject. Moreover, a subset of observers who demonstrate a high level of interobserver agreement can be identified by using pairwise agreement statistics between each observer and the internal majority standard opinion on each subject. These procedures are all illustrated within the context of a clinical diagnosis example involving seven pathologists.
Article
Controlled comparison of radiographic interpretive performance based on training and experience. This study compared each of these groups in medicine and chiropractic by testing abilities to interpret abnormal plain film radiographs of the lumbosacral spine and pelvis. Low back pain is a common and costly problem that is evaluated and treated primarily by medical physicians, orthopedists, and chiropractors. Although radiology is used extensively in patients with low back pain, the radiographic interpretations of students, clinicians, radiology residents, and radiologists have never been compared. Four hundred ninety-six eligible volunteers from nine target groups completed a test of radiographic interpretation consisting of nineteen cases with clinically important radiographic findings. The nine groups included 22 medical students, 183 chiropractic students, 27 medical radiology residents, 13 chiropractic radiology residents, 66 medical clinicians (including 12 general practice physicians, 25 orthopedic surgeons, 21 orthopedic residents, and 8 rheumatologists), 46 chiropractic clinicians, 48 general medical radiologists, 55 chiropractic radiologists, and 36 skeletal radiologists and fellows. The test established a high level of internal consistency reliability (0.880) and revealed that, in the interpretation of abnormal plain film radiographs of the lumbosacral spine and pelvis, significant differences were found among professional groups (P < 0.0001). Post hoc tests (P < 0.05) revealed that skeletal radiologists achieved significantly higher test results than did all other medical groups; that the test results of general medical radiologists and medical radiology residents was significantly higher than those of medical clinicians; that test results of medical students was significantly poorer than that of all other medical groups; that the performance of chiropractic radiologists and chiropractic radiology residents was significantly higher than that of chiropractic clinicians and chiropractic students; that no significant differences was revealed in the mean values of performance of chiropractic clinicians and chiropractic students; that the test results of chiropractic radiologists, chiropractic radiology residents, and chiropractic students was significantly higher than that of the corresponding medical categories (general medical radiologists, medical radiology residents, and medical students, respectively); that no significant difference in test results was identified between chiropractic radiologists and skeletal radiologists or between chiropractic and medical clinicians; and that the length of time in practice for clinicians and radiologists was not a significant factor in the test results. These data demonstrate a substantial increase in test results of all radiologists and radiology residents when compared to students and clinicians in both medicine and chiropractic related to the interpretation of abnormal radiographs of the lumbosacral spine and pelvis. Furthermore, the study reinforces the need for radiologic specialists to reduce missed diagnoses, misdiagnoses, and medicolegal complications.
Article
An intraobserver and interobserver study on the reproducibility of data was performed. This study investigates the variability in the interpretation of lumbar spine radiographs by chiropractors working in private practice. In chiropractic practice radiographs are used often, but this use is currently under debate. Therefore, there is a need for further study of the value of diagnoses made by radiographs by chiropractors. An acceptable intra- and interobserver agreement in radiograph reading is a prerequisite for a useful application of radiographs as a diagnostic tool in daily practice and in research. Four chiropractors read 100 blinded sets of standard, erect anteroposterior and lateral lumbar radiographs independently. The same set was read in two separate sessions with a 2-month interval. The first session revealed the interobserver agreement. The comparison of the ratings by the same assessor in the two sessions indicated the intraobserver agreement. The assessors used a specially developed criteria list with emphasis on "nonspecific" radiographic findings. The prevalence of some important categories was increased artificially. Agreement was expressed in percentage agreement and generalized kappa, combining the results of all four assessors. Most kappas ranged from 0.40 to 0.75, representing fair to good agreement. In general, intraobserver agreement was better than interobserver agreement. The low kappas that were found may be explained partially by the high-agreement-low-kappa paradox as a result of a low prevalence. The kappas and percentage agreement were acceptable, although not excellent. These results will be beneficial for future research on the value of radiograph diagnosis of nonspecific findings for delivery of safe and effective chiropractic therapy.
Article
An audit of Spanish private medicine radiodiagnostics facilities has been carried out, based partly on Spanish legislation relating to European Directives on health protection against ionizing radiation risks in medical exposure. The study included an appraisal of infrastructure and equipment, and aspects of quality assurance and radiation protection, by means of data collected through surveys. Of the 51 centres audited, a sample of 24 X-ray rooms was chosen, then an external evaluation with regard to image quality and patient dose was performed, by an advisory board of radiologists and medical physicists. The methodology used was similar to that of the group of European Union experts in European dose evaluation and image quality trials. Chest, abdomen, lumbar spine and breast examinations were monitored. Doses were measured with thermoluminescent dosimeters. A third of the X-ray rooms evaluated reached or exceeded dose reference values, and in a third of the cases the image quality left considerable room for improvement. Breast and chest examinations showed themselves to be the hardest to perform, not only as a result of exceeding the reference doses, but also due to failure to meet good image quality standards.
A study of pelvic radiography image quality in a Nigirian teaching hospital based on the Commission of European Communities (CEC) criteria
  • G B Inah
  • A O Akintomide
  • E E Edim
  • C Nzotta
  • N O Egbe
  • GB Inah
Inah GB, Akintomide AO, Edim EE, Nzotta C, Egbe NO. A study of pelvic radiography image quality in a Nigirian teaching hospital based on the Commission of European Communities (CEC) criteria. South Afr Radiogr. 2013;51:15-9.
Bekendtgørelse om medicinske røntgenanlæg til undersøgelse af patienter, nr. 975 af 16
  • Sundhedsstyrelsen
Sundhedsstyrelsen. Bekendtgørelse om medicinske røntgenanlaeg til undersøgelse af patienter, nr. 975 af 16. december 1998.
Nio 2MP (MDNC-2121) Specification sheet
  • Barco
Barco. Nio 2MP (MDNC-2121) Specification sheet, Article number K9601651. http://www.Barco.com. Belgium.
Interpretation of abnormal Lumbosakral spine radiographs -a test comparing students, clinicians, radiology residents and radiologists in medicine and chiropractic
  • Jam Taylor
  • P Clopton
  • E Bosch
  • K Miller
  • S Marcellis
Taylor JAM, Clopton P, Bosch E, Miller K, Marcellis S. Interpretation of abnormal Lumbosakral spine radiographs -a test comparing students, clinicians, radiology residents and radiologists in medicine and chiropractic. Spine. 1995;20:1147-53.