Top Banner
Postgraduate Medical Education and Training Board National Trainee Survey 2006 – key findings
60

National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

Sep 13, 2020

Download

Documents

dariahiddleston
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

Postgraduate Medical Educationand Training Board

National Trainee Survey 2006 – key findings

Page 2: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300
Page 3: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

Published by Postgraduate Medical Education and Training Board

Hercules HouseHercules RoadLondonSE1 7DU

Telephone: 020 7160 6100

www.pmetb.org.uk/traineesurvey

ISBN: 978-0-9555910-0-6

April 2007

Report written by Daniel Smith (PMETB), Patricia Le Rolland (PMETB) and ProfessorElisabeth Paice (London Deanery)

© PMETB 2007

National Trainee Survey 2006 – key findings 3

Page 4: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

4 National Trainee Survey 2006 – key findings

Contents

Acknowledgments..................................................................................6

Summary.............................................................................................. 7

Main findings....................................................................................................7

Supervision – the key to satisfaction.................................................................7

Layout of this report..........................................................................................8

1. Introduction.......................................................................................9

Background...................................................................................................... 9

Testing the face validity of the survey...............................................................9

Deaneries’ previous work............................................................................... 10

2. Methodology and data preparation...................................................... 11

Data collection................................................................................................11

Inclusion criteria.............................................................................................12

Data preparation.............................................................................................13

Response rates for the included cases............................................................13

The response rate and outlying indicator scores............................................ 14

Derivation of the indicator scores................................................................... 15

Method variance.............................................................................................25

Time in post....................................................................................................25

Adjustment for method variance.................................................................... 26

3. Key findings..................................................................................... 27

The survey’s outcome variables......................................................................27

Overall Satisfaction.........................................................................................28

Factors associated with job satisfaction.......................................................... 28

Medical Errors................................................................................................31

Supervision Score........................................................................................... 34

Feeling bullied – overall prevalence...............................................................35

The source of the perceived bullying............................................................. 35

Reporting of bullying and subsequent action................................................. 36

Are certain types of trainee more likely to be bullied?...................................36

Perceived bullying and indicator scores.........................................................38

National training initiatives: Radiology Academies and Hospital at Night...................................................................................... 39

Deaneries, training provider and differences in Overall Satisfactionand Supervision – a multilevel model analysis ...............................................44

Page 5: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

National Trainee Survey 2006 – key findings 5

4. The survey data as a quality management tool.......................................45

Data releases.................................................................................................. 45

The identification of outliers for follow up...................................................... 47

An example of identifying outliers – surgical specialty group........................ 47

Deanery quality improvement........................................................................ 48

5. The future of the National Trainee Survey..............................................50

Route of administration................................................................................... 50

The content of the survey................................................................................50

Adequate experience..................................................................................... 51

Overall Satisfaction Score............................................................................... 51

Follow up of concerns..................................................................................... 51

Personality variables...................................................................................... 51

Dissemination of the results............................................................................ 52

Population data...............................................................................................53

The survey as screening tool...........................................................................54

Making the survey mandatory.........................................................................55

Comparison with 2006 data.............................................................................55

National use of the data by Royal Colleges.....................................................55

The 2007 survey and other data sources.........................................................55

Deanery action plans......................................................................................56

Indicative plan for the 2007 survey................................................................. 56

Feedback form............................................................................................... 56

References.......................................................................................... 57

Appendices – available online from www.pmetb.org.uk/traineesurvey

Page 6: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

6 National Trainee Survey 2006 – key findings

Acknowledgments

PMETB thanks the following groups of people for their help with this work:

● all the deanery survey contacts who administered the survey in their area;

● all the postgraduate medical education centre staff who assisted with the datacollection;

● the deans from across the UK who supported this work.

The Surveys Working Group members:

Professor Elisabeth PaiceChair of COPMeD and Chair of Surveys Working Group

Maryanne AitkenCOPMeD National Trainee Survey Manager and Quality Manager, LondonDeanery

Dr Gellisse BagnallEducational Development Manager, NHS Education For Scotland

Dr Elaine DolmanTrent Deanery

Professor Shelley HeardPostgraduate Dean, London Deanery

Dr Stephen KellyAssociate Dean,West Midlands Deanery

Dr Mark RickenbachAssociate Dean, NHS Education South Central - Wessex Deanery

Dr Heather PayneAssociate Dean,Wales Deanery

Surveys Working Group PMETB staff:

Mark DexterHead of Policy

The contractors who assisted with data collection:

Nathan Collins WebLabs

Ray Flux CivilEyes

Bhavna Patel Document Capture Company

Samir Sayed Document Capture Company

Jane Smith Athene Communications

National Association of Medical Education Management

PMETB is responsible for this report’s contents.

Page 7: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

National Trainee Survey 2006 – key findings 7

Summary

Main findings

Overall the majority of trainees who took part responded positively to the itemsthey completed for the National Trainee Survey. This indicates that the majority oftrainees were broadly satisfied with their training posts but there were interestingvariations by provider and specialty. For example, it is possible to identifyspecialties and training providers that had low scores when compared to thenational figures and deaneries across the UK will use this information to target areasfor improvement. To ensure that the survey delivers improvements to trainees’experience of their training programmes, all UK deaneries have been using thedata from the survey to prepare action plans based on their results.

PMETB’s analysis has shown which facets of trainees’ experience of their posts wererelated to the survey’s outcome measures: these outcome measures were ‘OverallSatisfaction’ and ‘Medical Errors’. The analysis shows that better supervision isassociated with trainees reporting fewer perceived medical errors and greatersatisfaction with their posts.

Supervision – the key to satisfaction

The survey suggests that trainees’ perception of their satisfaction and supervisionvary by specialty grade group. GP trainees, for example, had the highest scores onboth these indicators, suggesting that they were both better supervised and moresatisfied than other trainees. Surgical group Senior House Officers (SHOs) had thelowest. Reported bullying also varied by specialty group and grade: it is interestingto note that the best supervised and most satisfied trainees – the GPs – had thelowest incidence of reported bullying.

The relationship between demographic variables and reported bullying varies byspecialty. As an example, female trainees were more likely to report being bulliedin anaesthetic, emergency medicine and surgical specialty groups than in others.

Survey data were also used to measure the impact of national initiatives on training.As examples, radiology trainees in academies reported higher Overall Satisfactionwith their posts than those not in academies and trainees working at sites in theHospital at Night initiative were more likely to report multidisciplinary handoversthan trainees who were not.

Page 8: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

8 National Trainee Survey 2006 – key findings

Layout of this report

Chapter 1 of this report covers the background to the survey. Chapter 2 covers thedata collection methodology and the derivation of the indicator scores used in thescreening tools that PMETB has provided to deaneries. Chapter 3 presents some ofthe key national findings from the data as a means to demonstrate that the indicatorscores are effective measures of providers’ adherence to PMETB’s Genericstandards for trainingi. Both chapters 2 and 3 contain statistical tables for interestedreaders. Chapter 4 gives details of how the indicator scores were used in thescreening tool. To avoid using the data as a league table and maintain the emphasison quality improvement, data on individual educational providers are not given.Deaneries do have indicator score data for providers in the screening tools.Chapter 5 makes recommendations for the next trainee survey based on PMETB’sevaluation of the 2006 work and feedback received from those who worked on theproject.

Throughout this report indicator scores are in title case, for instance OverallSupervision. This refers specifically to the Overall Satisfaction Score derived fromitems on this survey, as opposed to overall satisfaction generally.

The appendices referred to throughout are available online atwww.pmetb.org.uk/traineesurvey

Page 9: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

National Trainee Survey 2006 – key findings 9

1. Introduction

Background

In October 2005 PMETB inaugurated a Surveys Working Group chaired by ProfessorElisabeth Paice, charged with implementing a national survey of trainee doctors.This group reports to PMETB’s Statutory Training Committee. The stated purpose ofthe survey was to determine whether national training standards were being met.PMETB’s Generic standards for training were developed in tandem with items for theNational Trainee Survey and survey items were mapped to the standards (see Table5, Chapter 2). However, it was not possible to measure adherence to all of thestandards using a trainee survey instrument. Indeed, earlier work by Grant et alii

recognised that any trainee survey would need to be one part of a national data setused to quality assure postgraduate medical training. Grant et al proposed that thisnational data set could be used to inform quality assurance visits and to comparethe quality of training across providers and over time.

Grant et al noted that Professor Paice at London Deanery had developed a Point ofview survey to record trainees’ opinions about their trainingiii. They recommendedthat it would be appropriate to use London Deanery’s Point of view survey as thebasis for the proposed trainee instrument because:

● it addressed many of the issues included in the hospital visiting documentationused by Royal Colleges and deaneries;

● its core questions were developed and validated through face-to-face interviewswith over 300 SHOs;

● it had been successfully administered since 1996;

● the questions were suitable for every grade and specialty;

● it has also been used in Kent, Surrey Sussex (KSS) and Eastern deaneries.

PMETB followed this recommendation and the PMETB National Trainee Survey 2006used many items from the Point of view survey.

As part of the initial development of the survey, PMETB consulted with a number oftrainee groups, including the Academy of Medical Royal Colleges Trainee DoctorsGroup and the British Medical Association’s Junior Doctors Committee. Several ofthe Academy group’s suggestions were followed: adding an item on beingpressured to submit working hours that are compliant with the European WorkingTime Directive (item D2 on the survey) and an item on research opportunities, anddropping an item on trainees’ sexual orientation.

Testing the face validity of the survey

PMETB commissioned researchers at the University of Winchester to recruitvolunteer trainees, administer the survey to them and conduct interviews aftercompletion of the survey to ascertain that the face validity of items was acceptable.Data were obtained from 64 volunteers working in a range of specialties from fourdeaneries on:

● the clarity of the items and their understanding of the items;

● whether they had any problems understanding the item;

● whether the questionnaire covered all the relevant issues.

Page 10: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

10 National Trainee Survey 2006 – key findings

Participants’ comments were synthesised by the Winchester researchers intorecommendations for the surveyiv. It was possible to incorporate one of theserecommendations in the 2006 survey, namely the inclusion of items on makingmedical errors. The items used were derived from a United States study by Baldwinand Daugherty, which looked at the effect of sleep deprivation on residents’workingv. The Winchester group’s remaining recommendations will bereconsidered for the next National Trainee Survey.

Deaneries’ previous work

The majority of deaneries (17 out of the 18 for which there are data) had alreadyconducted surveys of trainees of one sort or another. PMETB’s and COPMeD’sintention continues to be that the National Trainee Survey would replace localsurveys at the time of administration to avoid questionnaire fatigue among traineesand to adhere to the principles of better regulation laid out in the concordatvi.Where deanery surveys occur more frequently than annually, their surveys that donot take place when the national survey is running can continue to take place.

Differences in methods of administering the previous surveys across deanerieswere apparent. Some deaneries chose to survey at a particular point in time(snapshot), while others chose to survey at the end of the trainees’ posts. Owing tothe fact that the PMETB and COPMeD National Trainee Survey needed to obtaindata on all posts within a limited timeframe, a snapshot approach was used; thismeans that respondents had been in post for varying lengths of time at the pointthey completed the survey. The Surveys Working Group will review the merits ofsnapshot versus end-of-post surveys when planning future survey work. It is worthnoting that the two approaches may not be mutually exclusive; for example, sometrainees could be surveyed at end of post during a snapshot survey.

Queries regarding the content of the report should be sent to:

[email protected]

Page 11: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

National Trainee Survey 2006 – key findings 11

2. Methodology and data preparation

Data collection

Data were collected between 15 May and 4 August 2006. Deaneries nominated asurvey contact for their deanery to work with the PMETB/COPMeD team. Deaneriesused one or more of the routes of administration detailed in Table 1 and werekindly supported by Postgraduate Medical Education Centres. Each route used thesame items, with minor variation in presentation where required by the format.

1 A summary of deaneries’ data collection is available here:http://www.pmetb.org.uk/fileadmin/user/QA/Trainee_Survey/PMETBCOPMEdTraineeSurveyBriefingNote3_1.pdf

2 Some emails did not get through the first time due to respondents having set their email accounts to haveenhanced junk mail filters. PMETB publicised this problem to ensure users checked their junk mail folders andresent them using an Outlook mail-merge rather than the database to get past the filter. Further consideration willbe given to this issue for future surveys.

Route Description Provider N (post-exclusions)

Portable electronicsurvey units(known as blackboxes)

PMETB –web by email

Respondents completedthe survey on theseunits by pressingnumeric keys. The unitswere placed inPostgraduate MedicalEducation Centres for alimited time-span.Postgraduate MedicalEducation Centre staffdirected the trainees tothe units.

PMETB set-up a website(www.traineesurvey.org.uk). Access to thesurvey was controlledby an individualpassword that wasemailed to the trainee.Reminder email wassent2.

Civil EyesResearchLimited

PMETB usingForms mastersoftware

6,463

Table 1 Routes of administration 1

PMETB –web by letter

PMETB set-up a website(www.traineesurvey.org.uk). Access to thesurvey was controlledby an individualpassword that was sentto the trainee by post asthe deanery wished touse the website but didnot have valid emailaddresses for thetrainees.

PMETB usingForms mastersoftware

7,488

Page 12: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

12 National Trainee Survey 2006 – key findings

Route Description Provider N (post-exclusions)

Data obtained through these routes were merged into one SPSS data file; this was aslow process as not all providers complied with the data template issued by PMETB.Final variables (i.e. the merged versions) were checked against the various sourcesto ensure that the data were not corrupted.

The items from the survey are given in Appendix 1 (available fromwww.pmetb.org.uk/traineesurvey).

PMETB –scannable paper form

The survey wasformatted to fit onto foursides of A4 (i.e. onepiece of A3 paper) forease of scanning. Aseparate sheet withlisting specialties andtheir codes wasprovided sorespondents couldanswer A5 and A6 usingcodes. Returns werescanned using OpticalMark and OpticalCharacter Recognition.

DocumentCaptureCompany

8,634

Severn andWessexwebsite

A website provided bythe suppliers of theIntrepid database andallowing the deanery tolink response directly totheir Intrepid data.

Hicomm 2,295

Inclusion criteria

The focus of the survey was trainees in educationally approved posts, so thefollowing criteria were used:

● Included: SHOs in approved posts, Specialist Registrars (SpRs), LocumAppointment Training posts (LATS), Fixed Term Training Appointments (FTTAs)and GP Registrars (GPRs).

● Excluded: Foundation Year 1 and Year 2 (F1s and F2s) (unless the deaneryincluded these for local analysis only)3, people in non-approved posts such asclinical fellows or trust SHOs, trainees on maternity leave, trainees on out ofprogramme experience (OOPE), non-medical public health specialists anddentists (unless dual-registered).

3 Following a request from the General Medical Council to avoid confusion with their work.

Page 13: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

National Trainee Survey 2006 – key findings 13

Grade N

Table 2 Grade exclusions

F1 1,669

F2 501

Trust doctor 551

Staff grade/other 739

Total 3,460

Data preparation

A total of 29,146 responses were received before the cut-off point for data entry. Thefollowing cases were excluded:

Grade exclusions

The questionnaire included response options for grade (item A1, see onlineAppendix 1) for groups of trainees who were excluded in case they received thesurvey by mistake and for foundation trainees included on a local basis only bysome deaneries. Table 2 gives the numbers of these respondents who were notincluded in PMETB reports.

Other exclusions

Cases were also excluded for the following reasons:

● data were missing on more than 14 items;

● there was a free text comment indicating that the respondent should not havebeen included, such as a reference to OOPE;

● there was a comment to say that the respondent was completing a second return.

This process left 24,880 cases available for analysis. The numbers available forindividual items on the survey may be below this due to “Not applicable” responseoptions and completion errors on the paper form.

Response rates for the included cases

As no master population file, listing all trainees by specialty and provider, wasavailable at the start of the survey, period data quality issues affected thecalculation of response rates. These were manifest in two ways: response ratesgreater than 100 per cent and locations with no denominator. In addition, recordingof SHO VTS (Vocational Training Scheme for GPs) was inconsistent on these returns.These problems were greater for finer aggregations, such as specialty groups,within a location. Given this, response rates are only reported by deanery here.Deaneries have received a response rate file that included the response rate foreach specialty group at each acute provider; 66 per cent (214 out of 326 acuteproviders with population data) had overall response rates of 50 per cent or more.

Page 14: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

14 National Trainee Survey 2006 – key findings

The response rate and outlying indicator scores

The provider level response file was merged with the aggregated data set and foreach specialty group correlations were obtained between:

● SHO response rate and the total number of outlying indicator scores below themean and indicator scores above the mean;

● SpR response rate and the total number of outlying indicator scores below themean and indicator scores above the mean.

The Bonferroni correctionvii was applied for multiple testing; only one of thecorrelations was found to be statistically significant at P <0.05. The surgery group:the correlation between the Surgery Group SHO response rate and the total numberof indicator scores lower than national mean (r = 0.27, P = 0.04). This suggests thatthe higher the response rate for this group of trainees at a provider the morenegative the indicator scores for the provider, although the correlation was weak.Overall, it can be concluded that providers’ response rates are not correlated withtheir indicator scores.

Table 3 Response rate by deanery

Deanery Response rate %

Eastern 100**Kent, Surrey, Sussex 68Leicestershire, Northamptonshire & Rutland 64London 64Mersey 52North Western 57Northern 66Northern Ireland 63Oxford 63Peninsula 65Scotland (East) 57Scotland (North) 97Scotland (South East) 47Scotland (West) 49Severn and Wessex 74South Yorkshire & South Humber 41Trent 56Wales 33West Midlands 81Yorkshire 65

**The response rate was greater than 100 per cent; presumably either sometrainees recorded their grades incorrectly or the population data were inaccurate.

Page 15: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

National Trainee Survey 2006 – key findings 15

Derivation of the indicator scores

Indicator scores were derived from the items by mapping to PMETB’s Genericstandards for training. This mapping is shown in Table 5 below, together with detailsof how the scores were calculated. For ordinal scores, the reliability coefficients4

measuring the score’s internal consistency are given, which, according toNunnally’sviii thresholds, are acceptable for the following scores: AdequateExperience, Feedback, Handover and Overall Satisfaction.

Items that were ordinal in nature (that is the responses that were on scales such as“Very Poor” to “Excellent”) were subjected to exploratory factor analysis5 to test forconstruct validityix. Items measuring a given construct should load6 on the samefactor and not on other factors. The results of the factor analysis are given in Table 4below.

Using Eigen values greater than 1 as a criterion, the factor analysis supports theconstruction of the following indicator scores, because the items used load on thesame factor and not on other factors, Overall Satisfaction, Supervision,Workload,Handover and Access to Resources.

However, other scales were more problematic; the Other Learning Opportunitiesand Feedback Score items did not load on their own factors and some of the itemsfrom these two scores did not load on any of the factors. The construction of both ofthese scores will be reviewed for the next survey.

The factor analysis failed to provide evidence of discriminant validity for theAdequate Experience Score; items D7 and D8 loaded on the same factor as theOverall Satisfaction Score items. This suggests that they are not measuring aseparate construct from Overall Satisfaction, perhaps because they are also overallrating items like the H2 to H6 items that are used to derive the Overall SatisfactionScore. PMETB plans to address this issue in 2007 using specialty specific items tolook at facets of experience. Rather than asking for a rating overall, the survey willask respondents about particular facets of their placement; for instance, there maybe items on theatre time for surgical trainees.

The results of the factor analysis were used to inform which items were used toderive the indicator scores but, due to the stated objective of the work, priority wasgiven to the mapping to the Generic standards for training when developing theindicator scores.

4 Cronbach’s alpha: the mean of all possible split-half reliability coefficients5 The goal of factor analysis is to summarise the pattern of correlations among the items to reduce the number of

items to a smaller number of factors.6 The loading is the correlation between the survey item and the factor, the size of the loading reflects the extent of

the relationship between the item and the factor.

Page 16: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

16 National Trainee Survey 2006 – key findings

Var

ian

ce e

xpla

ine

dH

4H

ow w

oul

d y

ou

rate

th

e q

uali

ty o

f exp

eri

en

ce in

th

is p

ost

?H

5H

ow w

oul

d y

ou

de

scri

be

th

is p

ost

to

a fr

ien

d w

ho

was

th

ink

ing

of a

pp

lyin

g f

or

it?

D7

How

wo

uld

yo

u ra

te t

he

pra

ctic

al e

xpe

rie

nce

yo

u ar

e g

ett

ing

in t

his

po

st?

H6

How

use

ful d

o y

ou

fee

l th

is p

ost

wil

l be

fo

r yo

ur fu

ture

car

ee

r?D

8H

ow c

on

fid

en

t ar

e y

ou

that

yo

ur c

urre

nt

po

st w

ill h

elp

yo

u ac

qui

re t

he

co

mp

ete

nce

s yo

u n

ee

d a

t th

is s

tag

e o

f yo

ur t

rain

ing

?H

3H

ow w

oul

d y

ou

rate

th

e q

uali

ty o

f sup

erv

isio

n in

th

is p

ost

?H

2H

ow w

oul

d y

ou

rate

th

e q

uali

ty o

f te

ach

ing

in t

his

po

st?

G7

In t

his

po

st,h

ow w

oul

d y

ou

rate

th

e e

nco

urag

em

en

t yo

u h

ave

had

to

tak

e s

tud

y le

ave

?C

2H

ow o

fte

n,i

f eve

r,h

ave

yo

u b

ee

n s

upe

rvis

ed

by

som

eo

ne

wh

o y

ou

fee

l isn

’t co

mp

ete

nt

to d

o s

o?

C1

How

oft

en

hav

e y

ou

felt

fo

rce

d t

o c

op

e w

ith

pro

ble

ms

bey

on

d y

our

co

mp

ete

nce

or

exp

eri

en

ce?

C4

Do

yo

u al

way

s k

now

wh

o is

pro

vid

ing

yo

ur c

lin

ical

sup

erv

isio

n w

he

n y

ou

are

wo

rkin

g?

C3

How

oft

en

hav

e y

ou

be

en

exp

ect

ed

to

ob

tain

co

nse

nt

for

pro

ced

ure

s w

hic

h y

ou

do

no

t ca

rry

out

yo

urse

lf?

C5

Ple

ase

ind

icat

e y

our

pe

rce

pti

on

of t

he

way

in w

hic

h c

riti

cal e

ven

ts a

nd

ne

ar m

isse

s ar

e r

ep

ort

ed

in y

our

de

par

tme

nt.

D9

How

wo

uld

yo

u ra

te t

he

inte

nsi

ty o

f yo

ur w

ork

,by

day

?D

10H

ow w

oul

d y

ou

rate

th

e in

ten

sity

of y

our

wo

rk,b

y n

igh

t?D

4H

ow o

fte

n h

as y

our

cur

ren

t w

ork

ing

pat

tern

left

yo

u fe

eli

ng

sh

ort

of s

lee

p w

he

n a

t w

ork

?D

3H

ow o

fte

n d

o y

ou

wo

rk b

eyo

nd

yo

ur r

ost

ere

d h

our

s?G

5*H

ow o

fte

n d

o y

ou

hav

e t

he

op

po

rtun

ity

to le

arn

to

ge

the

r w

ith

oth

er

he

alth

care

pro

fess

ion

als?

G3*

Ho

urs

of r

ele

van

t,ti

me

tab

led

,org

anis

ed

ed

ucat

ion

al m

ee

tin

gs

or

oth

er

eve

nts

of e

duc

atio

nal

val

ue o

n a

vera

ge

eac

h w

ee

kE

1H

ow o

fte

n h

ave

yo

u h

ad in

form

al f

ee

db

ack

fro

m a

se

nio

r cl

inic

ian

on

how

yo

u ar

e d

oin

g in

th

is p

ost

?H

1H

ow w

oul

d y

ou

rate

th

e q

uali

ty o

f in

duc

tio

n in

th

is p

ost

?D

5W

hic

h o

f th

e f

oll

owin

g b

est

de

scri

be

s h

and

ove

r ar

ran

ge

me

nts

AFT

ER

nig

ht

dut

y in

yo

ur p

ost

?D

6W

hic

h o

f th

e f

oll

owin

g b

est

de

scri

be

s h

and

ove

r ar

ran

ge

me

nts

BE

FOR

E n

igh

t d

uty

in y

our

po

st?

G9

Do

yo

u h

ave

acc

ess

to

th

e I

nte

rne

t at

yo

ur p

lace

of w

ork

?G

10H

ow e

asy

is it

to

ge

t ac

cess

to

th

e li

bra

ry s

erv

ice

s yo

u n

ee

d?

*Fo

r p

rese

nta

tio

n in

th

is t

able

th

e t

ext

of t

he

se it

em

s h

ave

be

en

sli

gh

tly

ed

ite

d.

Ext

ract

ion

Me

tho

d:P

rin

cip

al C

om

po

ne

nt

An

alys

is.R

ota

tio

n M

eth

od

:Var

imax

wit

h K

aise

r N

orm

alis

atio

n.R

ota

tio

n c

onv

erg

ed

in s

ix it

era

tio

ns.

On

ly lo

adin

gs

gre

ate

r th

an 0

.32

are

dis

pla

yed

.Kai

ser-

Mey

er-

Olk

in M

eas

ure

of S

amp

lin

g A

de

qua

cy =

0.8

97.N

= 1

6,60

9

19%

0.88

0.84

0.82

0.81

0.78

0.65

0.63

0.37

0.35

8% -0.6

6-0

.63

0.49

-0.4

70.

47

-0.3

3-0

.35

8% 0.76

0.73

0.62

0.50

7% 0.39

0.66

0.61

0.46

0.43

7% 0.89

0.88

6% 0.76

0.74

Q

Item

s

Overall Satisfaction

Supervision

Workload

Mixture -learning

Handover

Access toResources

Ta

ble

4

Fact

or a

na

lysi

s of

ord

ina

l su

rvey

ite

ms

Page 17: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

National Trainee Survey 2006 – key findings 17

Ta

ble

5

Th

e in

dic

ato

r sc

ores

Ind

ica

tor

scor

e

Acc

ess

to

Ed

ucat

ion

alR

eso

urce

s

Ad

eq

uate

Exp

eri

en

ce

Fee

db

ack

Do

mai

n 8

.Ed

ucat

ion

alre

sour

ces

and

cap

acit

y

8.2

Th

ere

mus

t b

e a

cce

ss t

oe

duc

atio

nal

fac

ilit

ies

(in

clud

ing

a li

bra

ry),

and

reso

urce

s (i

ncl

udin

g a

cce

ssto

th

e I

nte

rne

t in

all

wo

rkp

lace

s) o

f a s

tan

dar

d t

oe

nab

le t

rain

ee

s to

ach

ieve

th

eo

utco

me

s o

f th

e p

rog

ram

me

as s

pe

cifi

ed

in t

he

curr

icul

um.

G9.

Do

yo

u h

ave

acc

ess

to

the

In

tern

et

at y

our

pla

ce o

fw

ork

?

G10

.Hav

e y

ou

had

th

eo

pp

ort

unit

y to

par

tici

pat

e in

rese

arch

in t

his

po

st?

D7.

How

wo

uld

yo

u ra

te t

he

pra

ctic

al e

xpe

rie

nce

yo

u ar

eg

ett

ing

in t

his

po

st?

D8.

How

con

fid

en

t ar

e y

ou

that

yo

urcu

rre

nt

po

st w

ill h

elp

yo

uac

qui

re t

he

co

mp

ete

nce

syo

u n

ee

d a

t th

is s

tag

e o

fyo

ur t

rain

ing

?

E1.

How

oft

en

hav

e y

ou

had

info

rmal

fe

ed

bac

k fr

om

ase

nio

r cl

inic

ian

on

how

yo

uar

e d

oin

g in

th

is p

ost

? E

2.H

ave

yo

u h

ad a

fo

rmal

me

eti

ng

wit

h y

our

sup

erv

iso

r to

tal

k a

bo

ut y

our

pro

gre

ss in

th

is p

ost

? E

3.H

ave

yo

u h

ad f

orm

alas

sess

me

nt

of y

our

pe

rfo

rman

ce in

th

ew

ork

pla

ce?

Th

e it

em

s in

clud

ed

he

reun

de

r e

duc

atio

nal

re

sour

ces

are

th

ose

th

at r

ela

te t

oin

div

idua

l po

sts.

Lib

rary

serv

ice

s re

late

to

loca

tio

ns

as a

wh

ole

an

d a

re d

eal

tw

ith

els

ewh

ere

.

Pra

ctic

al e

xpe

rie

nce

is t

he

sine

qua

non

of p

ost

gra

dua

tetr

ain

ing

.Wh

ile

D7

and

D8

mig

ht

in t

he

ory

hav

e e

lici

ted

dif

fere

nt

resp

on

ses,

resp

on

ses

are

hig

hly

corr

ela

ted

.

Fee

db

ack

is a

n im

po

rtan

tfa

cto

r in

lear

nin

g.T

his

sco

reis

bas

ed

on

th

e a

vail

abil

ity

of d

ay-t

o-d

ay f

ee

db

ack

,ap

pra

isal

an

d a

sse

ssm

en

t.

Ite

ms

reco

de

d t

o 0

to

100

scal

e,w

he

re 1

00is

a g

oo

d s

core

.

A m

ean

is t

he

nca

lcul

ate

d o

r if

on

lyo

ne

ite

m is

pre

sen

t it

ssc

ore

is u

sed

.

0.41

0.80

0.72

Ite

ms

reco

de

d t

o 0

to

100

scal

e,w

he

re 1

00is

a g

oo

d s

core

.

A m

ean

is t

he

nca

lcul

ate

d,o

r if

on

lyo

ne

ite

m is

pre

sen

tit

s sc

ore

is u

sed

.

Ite

ms

reco

de

d t

o 0

to

100

scal

e,w

he

re 1

00is

a g

oo

d s

core

.

A m

ean

is t

he

nca

lcul

ate

d if

at

leas

ttw

o o

f th

e t

hre

e it

em

sar

e p

rese

nt.

5.1

Suff

icie

nt

pra

ctic

alex

pe

rie

nce

mus

t b

e a

vail

able

wit

hin

th

e p

rog

ram

me

to

sup

po

rt a

cqui

siti

on

of

com

pe

ten

ce a

s se

t o

ut in

th

ecu

rric

ulum

.

6.6

Tra

ine

es

mus

t h

ave

furt

he

r m

ee

tin

gs

wit

h t

he

ire

duc

atio

nal

sup

erv

iso

r (o

rre

pre

sen

tati

ve)

at le

ast

thre

e-

mo

nth

ly,t

o d

iscu

ss t

he

irp

rog

ress

,out

stan

din

gle

arn

ing

ne

ed

s an

d h

ow t

om

ee

t th

em

.

Do

mai

n 5

.De

live

ry o

fcu

rric

ulum

incl

udin

gas

sess

me

nt

Do

mai

n 6

.Sup

po

rtan

d d

eve

lop

me

nt

of

trai

ne

es,

trai

ne

rs a

nd

loca

l fac

ulty

PM

ET

B G

ener

icst

and

ard

s fo

r tr

ain

ing

Dom

ain

PM

ET

B G

ener

icst

an

da

rds

cove

red

Inte

rpre

tati

onIt

ems

incl

ud

edC

alc

ula

tion

Inte

rna

l re

lia

bil

ity

:C

ron

ba

ch’s

Alp

ha

(a)

Page 18: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

18 National Trainee Survey 2006 – key findings

Ind

ica

tor

Scor

e

Han

dov

er

Oth

erLe

arni

ngO

pp

ortu

nitie

s

Do

mai

n 1

.Pat

ien

tSa

fety

1.6

Tra

ine

es

in h

osp

ital

po

sts

mus

t h

ave

we

llo

rgan

ise

d h

and

ove

rar

ran

ge

me

nts

en

suri

ng

con

tinu

ity

of p

atie

nt

care

at t

he

sta

rt a

nd

en

d o

f day

or

nig

ht

dut

ies.

D5.

Wh

ich

of t

he

fo

llow

ing

be

st d

esc

rib

es

han

dov

er

arra

ng

em

en

ts B

EFO

RE

nig

ht

dut

y in

yo

ur p

ost

? D

6.W

hic

ho

f th

e f

oll

owin

g b

est

de

scri

be

s h

and

ove

rar

ran

ge

me

nts

AFT

ER

nig

ht

dut

y in

yo

ur p

ost

?

G1.

To w

hat

ext

en

t ar

e y

ou

invo

lve

d in

cli

nic

al a

udit

inth

is p

ost

? G

2.D

o y

ou

curr

en

tly

hav

e a

cce

ss t

o e

-le

arn

ing

mat

eri

al r

ele

van

t to

your

tra

inin

g?

G5.

How

oft

en

do

yo

u h

ave

th

e o

pp

ort

unit

yto

lear

n t

og

eth

er

wit

h o

the

rh

eal

thca

re p

rofe

ssio

nal

s(e

.g.n

urse

s,p

hysi

oth

era

pis

ts e

tc.)

? G

6.H

ave

yo

u ap

pli

ed

fo

r st

udy

leav

e in

th

is p

ost

? G

7 In

th

isp

ost

,how

wo

uld

yo

u ra

te t

he

en

cour

age

me

nt

you

hav

eh

ad t

o t

ake

stu

dy

leav

e?

G10

Hav

e y

ou

had

th

eo

pp

ort

unit

y to

par

tici

pat

e in

rese

arch

in t

his

po

st?

Hig

he

r sc

ore

s in

dic

ate

th

ath

and

ove

r is

mo

re f

orm

ally

org

anis

ed

an

d m

ore

like

ly t

ob

e in

clus

ive

of t

he

full

mul

ti-

pro

fess

ion

al t

eam

.

Th

is in

dic

ato

r co

mb

ine

s a

ran

ge

of u

nre

late

dad

dit

ion

al o

pp

ort

unit

ies.

Alo

w s

core

wo

uld

ind

icat

e t

he

ne

ed

to

exp

lore

wh

ich

of

the

se w

as p

rob

lem

atic

.

Ite

ms

reco

de

d t

o 0

to

100

scal

e,w

he

re 1

00is

a g

oo

d s

core

,so

alo

w s

core

ind

icat

es

ale

ss f

orm

al h

and

ove

r.

A m

ean

is t

he

nca

lcul

ate

d,o

r if

on

lyo

ne

ite

m is

pre

sen

t it

ssc

ore

is u

sed

.

0.78

0.36

Ite

ms

reco

de

d t

o 0

to

100

scal

e,w

he

re 1

00is

a g

oo

d s

core

.

A m

ean

is t

he

nca

lcul

ate

d if

at

leas

tfi

ve o

f th

e s

ix it

em

sar

e p

rese

nt.

6.13

Tra

ine

es

mus

tre

gul

arly

be

invo

lve

d in

th

ecl

inic

al a

udit

pro

cess

,in

clud

ing

pe

rso

nal

lyp

arti

cip

atin

g in

pla

nn

ing

,d

ata

coll

ect

ion

an

dan

alys

is.6

.17

Tra

ine

es

mus

th

ave

th

e o

pp

ort

unit

y to

lear

n w

ith

oth

er

he

alth

care

pro

fess

ion

als.

6.20

Tra

ine

es

mus

t b

e a

ble

to

tak

e s

tud

yle

ave

up

to

th

e m

axim

ump

erm

itte

d in

th

eir

te

rms

and

co

nd

itio

ns.

6.25

Tra

ine

es

sho

uld

be

exp

ose

d d

urin

g t

he

irtr

ain

ing

to

th

e a

cad

em

ico

pp

ort

unit

ies

avai

lab

le in

the

ir s

pe

cial

ty.

Do

mai

n 6

.Sup

po

rtan

d d

eve

lop

me

nt

of

trai

ne

es,

trai

ne

rs a

nd

loca

l fac

ulty

PM

ET

B G

ener

icst

and

ard

s fo

r tr

ain

ing

Dom

ain

PM

ET

B G

ener

icst

an

da

rds

cove

red

Inte

rpre

tati

onIt

ems

incl

ud

edC

alc

ula

tion

Inte

rna

l re

lia

bil

ity

:C

ron

ba

ch’s

Alp

ha

(a)

Page 19: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

National Trainee Survey 2006 – key findings 19

Ind

ica

tor

scor

e

Ove

rall

Sati

sfac

tio

nSc

ore

Sup

erv

isio

n

N/A

inte

rnal

out

com

em

eas

ure

N/A

inte

rnal

out

com

em

eas

ure

H2.

How

wo

uld

yo

u ra

te t

he

qua

lity

of t

eac

hin

g in

th

isp

ost

? H

3.H

ow w

oul

d y

ou

rate

th

e q

uali

ty o

fsu

pe

rvis

ion

in t

his

po

st?

H4.

How

wo

uld

yo

u ra

te t

he

qua

lity

of e

xpe

rie

nce

in t

his

po

st?

H5.

How

wo

uld

yo

ud

esc

rib

e t

his

po

st t

o a

frie

nd

wh

o w

as t

hin

kin

g o

fap

ply

ing

fo

r it

? H

6.H

owus

efu

l do

yo

u fe

el t

his

po

stw

ill b

e f

or

your

futu

reca

ree

r?

C1.

How

oft

en

hav

e y

ou

felt

forc

ed

to

co

pe

wit

hp

rob

lem

s b

eyo

nd

yo

urco

mp

ete

nce

or

exp

eri

en

ce?

C2.

How

oft

en

,if e

ver,

hav

eyo

u b

ee

n s

upe

rvis

ed

by

som

eo

ne

wh

o y

ou

fee

l isn

'tco

mp

ete

nt

to d

o s

o?

C3.

How

oft

en

hav

e y

ou

be

en

exp

ect

ed

to

ob

tain

con

sen

t fo

r p

roce

dur

es

wh

ich

yo

u d

o n

ot

carr

y o

utyo

urse

lf?

C4.

Do

yo

u al

way

s k

now

wh

ois

pro

vid

ing

yo

ur c

lin

ical

sup

erv

isio

n w

he

n y

ou

are

wo

rkin

g?

C5.

Ple

ase

ind

icat

e y

our

pe

rce

pti

on

of t

he

way

inw

hic

h c

riti

cal e

ven

ts a

nd

ne

ar m

isse

s ar

e r

ep

ort

ed

inyo

ur d

ep

artm

en

t.

Th

is in

dic

ato

r co

mb

ine

ssa

tisf

acti

on

wit

h e

ach

of t

he

key

ele

me

nts

of a

tra

inin

gp

ost

an

d p

rovi

de

s a

glo

bal

sati

sfac

tio

n s

core

.

Wh

ile

th

e q

uali

ty o

f pra

ctic

alex

pe

rie

nce

in a

po

st is

th

efa

cto

r m

ost

clo

sely

re

late

d t

oO

vera

ll S

atis

fact

ion

,th

eq

uali

ty o

f sup

erv

isio

n is

mo

st c

lose

ly r

ela

ted

to

th

ere

po

rtin

g o

f me

dic

al e

rro

rs.

Go

od

tra

inin

g r

eq

uire

s g

oo

dp

ract

ical

exp

eri

en

ce u

nd

er

safe

sup

erv

isio

n.

Ite

ms

reco

de

d t

o 0

to

100

scal

e,w

he

re 1

00is

a g

oo

d s

core

.

A m

ean

is t

he

nca

lcul

ate

d if

at

leas

tfo

ur o

f th

e f

ive

ite

ms

are

pre

sen

t.

0.89

0.54

Ite

ms

reco

de

d t

o 0

to

100

scal

e,w

he

re 1

00is

a g

oo

d s

core

.

A m

ean

is t

he

nca

lcul

ate

d if

at

leas

tfo

ur o

f th

e f

ive

ite

ms

are

pre

sen

t.

It c

ove

rs t

he

fo

llow

ing

man

dat

ory

sta

nd

ard

s:1.

1T

rain

ee

s m

ust

mak

e t

he

ne

ed

s o

f pat

ien

ts t

he

ir f

irst

con

cern

.1.2

Tra

ine

es

mus

t b

eap

pro

pri

ate

ly s

upe

rvis

ed

acco

rdin

g t

o t

he

ir e

xpe

rie

nce

and

co

mp

ete

nce

.1.3

Th

ose

sup

erv

isin

g t

he

cli

nic

al c

are

pro

vid

ed

by

trai

ne

es

mus

t b

ecl

ear

ly id

en

tifi

ed

,co

mp

ete

nt

to d

o s

o,ac

cess

ible

an

dap

pro

ach

able

by

day

an

d b

yn

igh

t,w

ith

tim

e f

or

the

sere

spo

nsi

bil

itie

s cl

ear

lyid

en

tifi

ed

wit

hin

th

eir

job

pla

n.1

.4 T

rain

ee

s m

ust

be

exp

ect

ed

to

ob

tain

co

nse

nt

on

ly f

or

pro

ced

ure

s w

hic

hth

ey a

re c

om

pe

ten

t to

pe

rfo

rm.

Do

mai

n 1

:Pat

ien

tSa

fety

PM

ET

B G

ener

icst

and

ard

s fo

r tr

ain

ing

Dom

ain

PM

ET

B G

ener

icst

an

da

rds

cove

red

Inte

rpre

tati

onIt

ems

incl

ud

edC

alc

ula

tion

Inte

rna

l re

lia

bil

ity

:C

ron

ba

ch’s

Alp

ha

(a)

Page 20: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

20 National Trainee Survey 2006 – key findings

Ind

ica

tor

scor

e

Wo

rkIn

ten

sity

Wo

rklo

ad

Do

mai

n 6

.Sup

po

rtan

d d

eve

lop

me

nt

of

trai

ne

es,

trai

ne

rs a

nd

loca

l fac

ulty

6.9

Wo

rkin

g p

atte

rns

and

inte

nsi

ty o

f wo

rk b

y d

ay a

nd

by

nig

ht

mus

t b

e a

pp

rop

riat

efo

r le

arn

ing

(n

eit

he

r to

o li

gh

tn

or

too

he

avy)

.

D9.

How

wo

uld

yo

u ra

te t

he

inte

nsi

ty o

f yo

ur w

ork

,by

day

? D

10.H

ow w

oul

d y

ou

rate

th

e in

ten

sity

of y

our

wo

rk,b

y n

igh

t?

D3.

How

oft

en

do

yo

u w

ork

bey

on

d y

our

ro

ste

red

ho

urs?

D4.

How

oft

en

has

yo

urcu

rre

nt

wo

rkin

g p

atte

rn le

ftyo

u fe

eli

ng

sh

ort

of s

lee

pw

he

n a

t w

ork

? D

9.H

oww

oul

d y

ou

rate

th

e in

ten

sity

of y

our

wo

rk,b

y d

ay?

D10

.H

ow w

oul

d y

ou

rate

th

ein

ten

sity

of y

our

wo

rk,b

yn

igh

t?

Th

is in

dic

ato

r m

ust

be

tre

ate

d w

ith

cau

tio

n:a

hig

hre

sult

is g

oo

d,b

ut a

low

resu

lt w

ill n

ot

dis

tin

gui

shb

etw

ee

n e

xce

ssiv

e o

rin

ade

qua

te w

ork

load

s.

Low

sco

res

are

an

ind

icat

or

of a

po

st w

he

re w

ork

inte

nsi

ty a

nd

/or

lon

g h

our

sm

ay le

ad t

o s

lee

pd

ep

riva

tio

n o

r ex

hau

stio

n.

Ite

ms

reco

de

d a

sfo

llow

s:0

'Ve

ryin

app

rop

riat

e w

ork

inte

nsi

ty -

to

o h

eav

yo

r to

o li

gh

t'

50 'S

lig

htl

yin

app

rop

riat

e w

ork

inte

nsi

ty -

sli

gh

tly

too

he

avy

or

slig

htl

y to

oli

gh

t'

100

'wo

rk in

ten

sity

abo

ut r

igh

t'

Th

en

th

e m

ean

of t

he

two

ite

ms

isca

lcul

ate

d,o

r if

on

lyo

ne

ite

m is

pre

sen

t it

ssc

ore

is u

sed

.

0.47

0.63

Ite

ms

reco

de

d t

o 0

to

100

scal

e,w

he

re 1

00is

a g

oo

d s

core

,so

alo

w s

core

is a

he

avy

wo

rklo

ad.

A m

ean

is t

he

nca

lcul

ate

d if

at

leas

tth

ree

of t

he

fo

ur it

em

sar

e p

rese

nt.

1.5

Shif

t an

d o

n-c

all r

ota

pat

tern

s m

ust

be

de

sig

ne

d s

oas

to

min

imis

e t

he

ad

vers

ee

ffe

cts

of s

lee

p d

ep

riva

tio

n.

Do

mai

n 1

:Pat

ien

tSa

fety

PM

ET

B G

ener

icst

and

ard

s fo

r tr

ain

ing

Dom

ain

PM

ET

B G

ener

icst

an

da

rds

cove

red

Inte

rpre

tati

onIt

ems

incl

ud

edC

alc

ula

tion

Inte

rna

l re

lia

bil

ity

:C

ron

ba

ch’s

Alp

ha

(a)

Page 21: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

National Trainee Survey 2006 – key findings 21

Ind

ica

tor

scor

e

Car

ee

rA

dvi

ce

Ho

urs

of

Ed

ucat

ion

Ed

ucat

ion

Sup

erv

isio

n

Do

mai

n 6

.Sup

po

rtan

d d

eve

lop

me

nt

of

trai

ne

es,

trai

ne

rs a

nd

loca

l fac

ulty

6.8

Th

ere

mus

t b

e r

ead

yac

cess

to

car

ee

r ad

vice

.E

4.H

ave

yo

u h

ad a

dis

cuss

ion

wit

h a

se

nio

rco

lle

ague

ab

out

yo

ur c

are

er

pla

ns?

G3.

How

man

y h

our

s o

fre

leva

nt,

tim

eta

ble

d,

org

anis

ed

ed

ucat

ion

alm

ee

tin

gs

or

oth

er

eve

nts

of

ed

ucat

ion

al v

alue

do

yo

uta

ke p

art

in o

n a

vera

ge

eac

hw

ee

k?

F1.D

o y

ou

hav

e a

de

sig

nat

ed

ed

ucat

ion

alsu

pe

rvis

or?

F2.

Do

yo

u h

ave

a tr

ain

ing

/le

arn

ing

agre

em

en

t w

ith

yo

ursu

pe

rvis

or,

sett

ing

out

yo

urre

spe

ctiv

e r

esp

on

sib

ilit

ies?

F3.A

re y

ou

usin

g a

lear

nin

gp

ort

foli

o in

th

is p

ost

? F4

.Are

you

usin

g a

log

bo

ok

in t

his

po

st?

F5.H

ave

yo

u b

ee

n t

old

wh

om

to

tal

k t

o in

con

fid

en

ce if

yo

u h

ave

con

cern

s,p

ers

on

al o

re

duc

atio

nal

?

Car

ee

r ad

vice

is p

arti

cula

rly

imp

ort

ant

for

the

SH

Og

rad

e.T

his

ind

icat

or

is n

ot

com

par

ed

to

th

e n

atio

nal

me

an o

r in

clud

ed

in t

he

tota

ls b

elo

w a

nd

ab

ove

th

en

atio

nal

me

an,b

eca

use

con

fid

en

ce in

terv

als

can

no

tb

e c

alcu

late

d a

s fo

r so

me

loca

tio

ns

the

sta

nd

ard

dev

iati

on

was

0.

Th

is in

dic

ato

r lo

ok

s at

th

eh

our

s o

f we

ek

ly e

duc

atio

n.

Tra

ine

es

are

un

like

ly t

o t

ake

into

acc

oun

t m

on

thly

or

less

fre

que

nt

reg

ion

al t

rain

ing

day

s,e

tc.

Th

is in

dic

ato

r is

ab

out

th

ee

duc

atio

nal

fram

ewo

rkun

de

rpin

nin

g t

he

po

st.

Eve

ry e

lem

en

t o

f th

efr

amew

ork

is a

sso

ciat

ed

wit

hg

oo

d t

rain

ing

.Low

sco

res

sug

ge

st a

tte

nti

on

sh

oul

d b

ep

aid

to

pro

gra

mm

em

anag

em

en

t an

d t

he

ro

le o

fth

e d

ire

cto

r o

f me

dic

ale

duc

atio

n in

en

suri

ng

stru

ctur

es

and

sys

tem

s ar

ein

pla

ce.

Ite

ms

reco

de

d t

o 0

to

100

scal

e,w

he

re 1

00is

a g

oo

d s

core

.

n/a

n/a

n/a

Me

an n

umb

er

of

ho

urs.

Th

ere

is a

ceil

ing

of e

igh

t,w

hic

his

lab

ell

ed

8 o

r m

ore

.

Sum

of "

Yes"

resp

on

ses

acro

ss t

he

five

ite

ms.

Sco

re c

anb

e fr

om

0 t

o 5

.If a

nit

em

has

no

re

spo

nse

,n

o s

core

isca

lcul

ate

d.

5.3

Tra

ine

es

mus

t b

e a

ble

to

acce

ss a

nd

be

fre

e t

o a

tte

nd

trai

nin

g d

ays,

cour

ses

and

oth

er

mat

eri

al t

hat

fo

rms

anin

trin

sic

par

t o

f th

e t

rain

ing

pro

gra

mm

e.

6.3

Tra

ine

es

mus

t h

ave

ad

esi

gn

ate

d e

duc

atio

nal

sup

erv

iso

r.6.

4 T

rain

ee

s m

ust

sig

n a

tra

inin

g/l

ear

nin

gag

ree

me

nt

at t

he

sta

rt o

f eac

hp

ost

.6.5

Tra

ine

es

mus

t h

ave

alo

gb

oo

k a

nd

/or

a le

arn

ing

po

rtfo

lio

re

leva

nt

to t

he

ircu

rre

nt

pro

gra

mm

e,w

hic

hth

ey d

iscu

ss w

ith

th

eir

ed

ucat

ion

al s

upe

rvis

or

(or

rep

rese

nta

tive

).6.

7 T

rain

ee

sm

ust

hav

e a

me

ans

of f

ee

din

gb

ack

in c

on

fid

en

ce t

he

irco

nce

rns

and

vie

ws

abo

utth

eir

tra

inin

g a

nd

ed

ucat

ion

exp

eri

en

ce t

o a

n a

pp

rop

riat

em

em

be

r o

f lo

cal f

acul

ty.

Do

mai

n 5

.De

live

ry o

fcu

rric

ulum

incl

udin

gas

sess

me

nt

Do

mai

n 6

.Sup

po

rtan

d d

eve

lop

me

nt

of

trai

ne

es,

trai

ne

rs a

nd

loca

l fac

ulty

PM

ET

B G

ener

icst

and

ard

s fo

r tr

ain

ing

Dom

ain

PM

ET

B G

ener

icst

an

da

rds

cove

red

Inte

rpre

tati

onIt

ems

incl

ud

edC

alc

ula

tion

Inte

rna

l re

lia

bil

ity

:C

ron

ba

ch’s

Alp

ha

(a)

Page 22: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

22 National Trainee Survey 2006 – key findings

Ind

ica

tor

scor

e

Ind

ucti

on

Bul

lyin

g b

yC

on

sult

ants

6.1

Eve

ry t

rain

ee

sta

rtin

g a

po

st o

r p

rog

ram

me

mus

tat

ten

d a

de

par

tme

nta

lin

duc

tio

n t

o e

nsu

re t

hey

und

ers

tan

d t

he

cur

ricu

lum

,h

ow t

he

ir p

ost

fit

s w

ith

in t

he

pro

gra

mm

e,th

eir

dut

ies

and

rep

ort

ing

arr

ang

em

en

ts,t

oe

nsu

re t

hey

are

to

ld a

bo

utd

ep

artm

en

tal p

oli

cie

s an

d t

om

ee

t ke

y st

aff.

6.2

At

the

sta

rt o

f eve

ry p

ost

wit

hin

a p

rog

ram

me,

the

ed

ucat

ion

al s

upe

rvis

or

(or

rep

rese

nta

tive

) m

ust

dis

cuss

wit

h t

he

tra

ine

e t

he

ed

ucat

ion

al fr

amew

ork

an

dsu

pp

ort

sys

tem

s in

th

e p

ost

and

th

e r

esp

ect

ive

resp

on

sib

ilit

ies

of t

rain

ee

an

dtr

ain

er

for

lear

nin

g.T

his

dis

cuss

ion

sh

oul

d in

clud

e t

he

sett

ing

of a

ims

and

ob

ject

ive

sfo

r th

e t

rain

ee

to

ach

ieve

inth

e p

ost

.

B1

Did

so

me

on

e e

xpla

inyo

ur r

ole

an

dre

spo

nsi

bil

itie

s in

yo

ur u

nit

or

de

par

tme

nt

at t

he

sta

rt o

fth

is p

ost

? B

2 D

id y

ou

ge

t al

lth

e in

form

atio

n y

ou

ne

ed

ed

abo

ut y

our

wo

rkp

lace

wh

en

you

star

ted

wo

rkin

g t

he

re?

B3.

Did

yo

u si

t d

own

wit

hyo

ur s

upe

rvis

or

and

dis

cuss

your

ed

ucat

ion

al o

bje

ctiv

es

for

your

cur

ren

t p

ost

?

A g

oo

d in

duc

tio

n s

ets

th

eto

ne

fo

r th

e w

ho

le p

ost

.

n/a

n/a

Sum

of "

Yes"

resp

on

ses

acro

ss t

he

five

ite

ms.

Sco

re c

anb

e fr

om

0 t

o 3

.If a

nit

em

has

no

re

spo

nse

,n

o s

core

isca

lcul

ate

d.

6.11

Tra

ine

es

mus

t n

ot

be

sub

ject

ed

to,

or

sub

ject

oth

ers

to,

be

hav

iour

th

atun

de

rmin

es

the

irp

rofe

ssio

nal

co

nfi

de

nce

or

self

-est

ee

m.

J1.H

ave

yo

u b

ee

nsu

bje

cte

d t

o p

ers

iste

nt

be

hav

iour

in t

his

po

st t

hat

has

un

de

rmin

ed

yo

urp

rofe

ssio

nal

co

nfi

de

nce

and

se

lf-e

ste

em

?

J3.W

hic

h o

ne

of t

he

foll

owin

g is

th

e m

ain

sour

ce o

f th

is b

eh

avio

ur?

Perc

en

tag

e o

f re

spo

nd

en

tsw

ho

re

po

rt b

ein

gsu

bje

cte

d t

o b

eh

avio

urd

efi

ne

d in

J1

by

con

sult

ants

- t

he

irre

spo

nse

to

J3.

Bul

lyin

g m

ay a

lso

be

pe

rpe

trat

ed

by

oth

er

trai

ne

es,

nurs

es,

man

age

rsan

d e

ven

pat

ien

ts.

Th

is in

dic

ato

r o

nly

loo

ks

at c

on

sult

ant

bul

lyin

g,b

eca

use

this

so

urce

isas

soci

ate

d w

ith

th

em

ost

str

ess

Do

mai

n 6

.Sup

po

rtan

d d

eve

lop

me

nt

of

trai

ne

es,

trai

ne

rs a

nd

loca

l fac

ulty

Do

mai

n 6

.Sup

po

rtan

d d

eve

lop

me

nt

of

trai

ne

es,

trai

ne

rs a

nd

loca

l fac

ulty

PM

ET

B G

ener

icst

and

ard

s fo

r tr

ain

ing

Dom

ain

PM

ET

B G

ener

icst

an

da

rds

cove

red

Inte

rpre

tati

onIt

ems

incl

ud

edC

alc

ula

tion

Inte

rna

l re

lia

bil

ity

:C

ron

ba

ch’s

Alp

ha

(a)

Page 23: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

National Trainee Survey 2006 – key findings 23

Ind

ica

tor

scor

e

Me

dic

alE

rro

r

Eur

op

ean

Wo

rkin

gT

ime

Dir

ect

ive

pe

rce

nta

ge

Do

mai

n 2

.Qua

lity

Ass

uran

ce,R

evie

wan

d E

valu

atio

n

2.1

Pro

gra

mm

es,

po

sts,

asso

ciat

ed

man

age

me

nt,

and

dat

a co

lle

ctio

n c

on

cern

ing

trai

ne

es

and

loca

l fac

ulty

mus

t co

mp

ly w

ith

th

eE

uro

pe

an W

ork

ing

Tim

eD

ire

ctiv

e,D

ata

Pro

tect

ion

Act

and

Fre

ed

om

of I

nfo

rmat

ion

Act

.

N/A

inte

rnal

out

com

em

eas

ure

to

ass

ess

th

ep

roce

ss s

tan

dar

ds

rela

tin

g t

op

atie

nt

safe

ty in

do

mai

n 1

of

the

Gen

eric

sta

ndar

ds

for

trai

ning

.

D1.

Are

yo

ur r

ost

ere

dw

ork

ing

ho

urs

com

pli

ant

wit

h t

he

Eur

op

ean

Wo

rkin

gT

ime

Dir

ect

ive

?

D2.

Hav

e y

ou

be

en

ask

ed

to

sub

mit

ho

urs

that

are

com

pli

ant

wit

h t

he

Eur

op

ean

Wo

rkin

g T

ime

Dir

ect

ive,

wh

en

th

e h

our

s yo

u ac

tual

lyw

ork

are

no

t co

mp

lian

t

H7

In t

he

last

mo

nth

hav

eyo

u m

ade

a s

eri

ous

me

dic

ale

rro

r?

H8

In t

he

last

mo

nth

,hav

eyo

u m

ade

a p

ote

nti

ally

seri

ous

me

dic

al e

rro

r?

Th

is s

core

is d

iffi

cult

to

inte

rpre

t,as

it is

no

tm

eas

urin

g a

ctua

l err

ors

but

wh

eth

er

the

tra

ine

e w

ill

rep

ort

hav

ing

mad

e t

he

m o

nth

is s

urve

y in

stru

me

nt.

Perc

en

tag

e o

fre

spo

nd

en

ts w

ho

answ

ere

d "

Yes"

to

D1

and

"N

o"

to D

2.If

an

ite

m h

as n

o r

esp

on

se,

no

sco

re is

calc

ulat

ed

.

n/a

n/a

Sco

red

1 if

th

ere

spo

nd

en

t in

dic

ate

dth

ey h

ad m

ade

on

e o

rm

ore

th

an o

ne

seri

ous

an

d/o

rp

ote

nti

ally

se

rio

usm

ed

ical

err

or

in t

he

last

mo

nth

.Oth

erw

ise

sco

red

as

0.R

esp

on

de

nts

wit

hm

issi

ng

dat

a o

r w

ho

ind

icat

ed

th

ey d

id n

ot

wis

h t

o a

nsw

er

we

rese

t to

mis

sin

g a

nd

excl

ude

d fr

om

th

ean

alys

is.

N/A

inte

rnal

out

com

em

eas

ure

PM

ET

B G

ener

icst

and

ard

s fo

r tr

ain

ing

Dom

ain

PM

ET

B G

ener

icst

an

da

rds

cove

red

Inte

rpre

tati

onIt

ems

incl

ud

edC

alc

ula

tion

Inte

rna

l re

lia

bil

ity

:C

ron

ba

ch’s

Alp

ha

(a)

Page 24: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

24 National Trainee Survey 2006 – key findings

Ind

ica

tor

scor

e

Oth

er

Form

alTe

ach

ing

6.16

Tra

ine

es

mus

t b

e a

ble

to

acce

ss t

rain

ing

in g

en

eri

cp

rofe

ssio

nal

sk

ills

at

all

stag

es

in t

he

ir d

eve

lop

me

nt.

G4.

Hav

e y

ou

had

fo

rmal

teac

hin

g s

ince

leav

ing

me

dic

al s

cho

ol i

n a

ny o

f th

efo

llow

ing

?

Co

mm

unic

atio

n,a

pp

rais

alsk

ills

,te

ach

ing

sk

ills

,le

ade

rsh

ip,t

eam

wo

rkin

g,

pat

ien

t sa

fety

,tim

em

anag

em

en

t,m

ed

ical

eth

ics.

As

the

wo

rdin

g o

f th

ese

ite

ms

did

no

t p

ert

ain

to

th

ecu

rre

nt

po

st,t

his

ind

icat

or

can

on

ly b

e a

pp

lie

d a

td

ean

ery

leve

l an

d n

ot

the

leve

l of p

rovi

de

r.

Sco

red

1 f

or

eac

hg

en

eri

c sk

ill t

he

trai

ne

e h

as r

ece

ive

dtr

ain

ing

in.P

oss

ible

ran

ge

0 t

o 8

.

n/a

PM

ET

B G

ener

icst

and

ard

s fo

r tr

ain

ing

Dom

ain

PM

ET

B G

ener

icst

an

da

rds

cove

red

Inte

rpre

tati

onIt

ems

incl

ud

edC

alc

ula

tion

Inte

rna

l re

lia

bil

ity

:C

ron

ba

ch’s

Alp

ha

(a)

Page 25: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

National Trainee Survey 2006 – key findings 25

Method variance

Route of administration

The data were analysed to test for any differences by route of administration (seeTable 1 above). Differences did emerge. Respondents who submitted a return usingthe portable electronic survey units seemed more likely to give more negativeresponses. Some examples of this effect are given below in Table 6 for items D2 andH7. It seems likely that this effect is due to trainees perceiving the survey units asmore anonymous than the web (trainees were contacted by email) and paper (barcodes were present on the paper forms), so respondents using the survey unitswere less inclined to give more socially desirable responses as they were confidentthey could not be identified.

Time in post

As the survey was a snapshot, respondents had been in post for variable lengths oftime. Comparing the indicator scores against the time in post item (A2), showed thattrainees who had been in post longer were slightly more likely to give morepositive answers.

All other routes

In the last month, haveyou made a seriousmedical error? -percentage yes to onceor more than once

Have you been asked tosubmit hours that arecompliant with theEuropean Working TimeDirective, when thehours you actually workare NOT compliant? -percentage yes

Portableelectronic surveyunits

Portableelectronic surveyunits

All other routes

2.0

5.5

11.9

33.5

18,102

6,365

16,563

201.98

1,332.77

<0.001

<0.001

5,342

N x2 p

Table 6 Examples of method variance

Page 26: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

26 National Trainee Survey 2006 – key findings

Adjustment for method variance

This method variance was particularly problematic because route of administrationwas confounded by provider and deanery. In order to make valid comparisonsacross providers or deaneries and be certain that differences were not due tovariation in the route or the length of time in post, all the indicator scores8 wereadjusted as follows:

Each indicator score was regressed on to the survey route recoded (1 for portableelectronic survey units and 0 for all other routes) and the length of time in post (asper A2).

An adjusted indicator score was calculated by applying the regression coefficientsobtained in step 1: adjusted score = score + (survey route recoded * B1) + (lengthof time in post * B2).

There were then no differences in mean adjusted indicator scores by route or timein post.

It was not possible to adjust the indicator scores based on categorical items at therespondent level using this method. So these should be interpreted in isolationrather than comparing across providers.

8 Access to Educational Resources, Adequate Experience, Feedback, Handover, Other Learning Opportunities,Overall Satisfaction, Supervision,Work Intensity and Workload

Page 27: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

National Trainee Survey 2006 – key findings 27

3. Key findings

The objective of the analysis presented in this chapter is to demonstrate that theindicator scores derived from the survey have both concurrent and constructvalidityix and that the scores can show differences across providers in trainees’perceptions of the quality of their training. The successful demonstration of thesefeatures of the scores shows that the survey can be appropriately used as ascreening tool of quality assurance and quality management work. To demonstrateconcurrent validity, analysis was undertaken using the surveys outcome variables.To demonstrate construct validity, the survey was used to assess national traininginitiatives. Finally, analysis was undertaken to look for differences across trainingproviders on the indicator scores that earlier analysis had shown to be the mostvalid and reliable of the scores.

The survey’s outcome variables

The survey collected a number of items that may be viewed as outcome variables:items used in the Overall Satisfaction Score and the Medical Error Score; andwhether the trainee experienced behaviour that undermined their self-confidence(colloquially this might be termed bullying). Trainees’ responses to these outcomevariables can be predicted using the other scores from the survey to answerquestions such as which aspects of trainees’ perceptions of their posts are moststrongly related to trainees’ Overall Satisfaction Scores. If relationships betweenthese measures of aspects of the trainees’ experience and these outcomes variablesare found, they provide evidence that the indicator scores have concurrent validity;i.e. trainees’ perceptions of facets of their training relate to outcomes measures thatwere obtained at the same time as the facet measures, such as the OverallSatisfaction Score. One would expect these relationships to exist, so the object is notto demonstrate that the relationships exist (as one might do in research), but toprovide evidence that the survey items are measuring what they purport to. Forexample, one would expect the supervision of trainees to be related to whetherthey report having made perceived medical errors; therefore the Supervision Scoreshould be associated with the Medical Error Score.

Page 28: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

90

GP (N =

1,856)

Patholo

gy SHO (N =

96)

Psych

iatry

SpR (N =

773)

Opthalm

ology SpR (N

= 267)

Anaesthetic

s SHO (N

= 1,1

86)

Opthalm

ology SHO (N

= 194)

Emerg

ency M

edicin

e SHO (N =

902)

Radiolo

gy SpR (N =

818)

Psych

iatry

SHO (N =

1,674)

Surgery

SpR (N =

2,101)

Patholo

gy SpR (N =

467)

Paediatri

cs and C

hild H

ealth SHO (N

= 1,2

14)

Medic

ine SpR (N

= 2,9

83)

Anaesthetic

s SpR (N

= 1,4

19)

Paediatri

cs and C

hild H

ealth SpR (N

= 1,1

13)

Emerg

ency M

edicin

e SpR (N =

350)

Obstetri

cs and G

ynaecolo

gy SpR (N =

882)

Obstetri

cs and G

ynaecolo

gy SHO (N =

906)

Medic

ine SHO (N

= 2, 9

05)

Surgery

SHO (N =

1,905)

Overall Satisfaction

Ove

rall

Sat

isfa

ctio

n

Sco

re A

dju

ste

d

85

80

75

70

65

60

55

50

28 National Trainee Survey 2006 – key findings

On this measure, surgical SHOs are the least satisfied with their training. Looking atthe items included in the Overall Satisfaction Score, 3.9 per cent of surgical SHOs (N= 1,904) indicated that they would describe their post as very poor to a friend,compared with only 0.4 per cent of GPRs (N = 1,853). Similarly, 1.6 per cent (N =1,907) of SHO surgeons rated their supervision as very poor compared with only0.2 per cent (N =1,855) of the GPs. In terms of numbers of trainees, there are 1,237trainees in the specialties and grades on Chart 1 (5.2 per cent) who indicated thatthey would describe their post as poor or very poor to a friend.

Overall Satisfaction

Trainees’ Overall Satisfaction with their training posts varies by specialty andgrade; this is illustrated in Chart 19.

Chart 1 Trainees’ Overall Satisfaction Score by specialty/grade groups

Factors associated with job satisfaction

The two scores most closely associated with trainees’ Overall Satisfaction Score areSupervision and Adequate Experience; good posts are those where trainees areexposed to useful experience under good supervision. However, the AdequateExperience Score was excluded from the analysis that follows because the factoranalysis (see Table 4, Chapter 2) suggested that it is not measuring a separateconstruct from the Overall Satisfaction Score. The two scores are highly correlated(r = 0.759, N = 24,843). This measurement problem may be partly due toperceptions of the adequacy of the experience gained being so inextricably boundup in trainees’ overall satisfaction with post; it is not possible to be satisfied with apost that provides inadequate experience. Another plausible explanation,mentioned in Chapter 2, is that the Adequate Experience items are also overallrating type items like those used in the Overall Satisfaction Score. Plans to tacklethis measurement issue are outlined in Chapter 5.

9 The chart excludes Public Health and Occupational Medicine due to wide confidence intervals. GPs includesGPRs and SHO based in GP practices, but not SHO VTS trainees based in acute sites who are included in theappropriate acute specialty group.

Page 29: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

National Trainee Survey 2006 – key findings 29

Sequential multiple regression was used to examine which facets of the trainees’job, as measured by the survey’s indicator scores, were most strongly related totheir Overall Satisfaction Score. Independent variables were entered into the modelin the order given in Table 7 below. At each step, the addition of the new variablessignificantly improved the prediction. A considerable amount of the variance (39per cent) in the Overall Satisfaction Score is explained by the survey indicatorscores entered in step three of the multiple regression. This is over and above thevariance accounted for by the variables entered in steps one and two. So, forinstance, one could not argue that difference by specialty group on the indicatorscores are the explanation for differences in the Overall Satisfaction Score, becausethe model has already accounted for differences by specialty group.

Table 7 Overall model predicting the trainees’ Overall Satisfaction with theirtraining

10 Ethnicity was recoded into white or non-white for the purposes of this analysis.11 Derived from A5 and A6, e.g. respondent working in medicine and intending to practise as consultant in surgery.

Demographicvariables

Time in current postSurvey route

0.008 0.008 86.465 <0.001

0.074 0.073 81.489 <0.001

0.396 0.395 1,439.848 <0.001

Method variance1

2

3 Survey indicatorscores

N = 21,660. All specialties included

GradeSpecialty group (dummy coded)Ethnicity10

SexDifferent group from intended11

Forces Year qualifiedWhere qualified

Access to EducationalResources Bullying by ConsultantsEducational SupervisionHours of Education Induction Other Learning Opportunities Supervision Workload

R2Step Type of variables Variables R2adj F change P

Page 30: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

30 National Trainee Survey 2006 – key findings

The following indicator scores were excluded from this model:

● Career Advice, Feedback, Handover and European Working Time scores. Thesewere excluded because the number of not applicable/not sure/missing responsesreduced the number of cases available for the analysis (all had over 3,650respondents with missing data).

● The Work Intensity Score was removed because it contains items (D9 and D10)that were also included in the Workload Score (though coded differently) and istherefore not independent.

Table 8 Survey variables and Overall Satisfaction

2.015

6.838

-0.632

-0.004

-0.392

-2.962

-3.403

-0.320

-5.325

-9.673

-5.936

-7.374

-4.302

-4.752

-3.241

-4.366

-5.548

-1.029

-1.468

3.325

-7.957

-7.762

9.697

-14.426

15.583

21.923

23.718

33.856

49.129

-1.321

0.044

0.000

0.527

0.996

0.695

0.003

0.001

0.749

0.000

0.000

0.000

0.000

0.000

0.000

0.001

0.000

0.000

0.303

0.142

0.001

0.000

0.000

0.000

0.000

0.000

0.000

0.000

0.000

0.000

0.187

0.001

0.000

0.000

0.000

0.000

0.000

0.000

0.001

0.003

0.001

0.002

0.001

0.001

0.000

0.001

0.001

0.000

0.000

0.000

0.002

0.002

0.003

0.006

0.007

0.013

0.016

0.032

0.067

0.000

Constant

Survey route

Time in post

Grade (SpR/GPR = 1)

Forces (Forces =1)

Anaesthetics

Emergency medicine

General practice

Medicine

Obstetrics and gynaecology

Occupational medicine

Paediatrics and child health

Pathology

Psychiatry

Public health

Radiology

Surgery

Different group from intended

Year qualified

Sex (female = 1)

Ethnicity (minority ethnic group = 1)

Where qualified (outside UK = 1)

Access to Educational Resources Score

Bullying by Consultants Score (1 = reported)

Educational Supervision Score

Hours of Education Score

Induction Score

Other Learning Opportunities Score

Supervision Score

Workload Score

Shared variance = 0.238, Unique variance = 0.158

Variables that are statistically significant at P < 0.05 are shaded.

tPredictor

0.038

-0.004

-0.000

-0.002

-0.038

-0.034

-0.004

-0.091

-0.109

-0.034

-0.092

-0.033

-0.059

-0.020

-0.039

-0.083

-0.006

-0.010

0.019

-0.051

-0.055

0.055

-0.078

0.095

0.128

0.145

0.209

0.310

-0.008

B Standardised

Sig. sr2

97.448

1.283

-0.074

-0.001

-0.357

-1.866

-2.357

-0.208

-3.194

-6.329

-9.222

-4.687

-3.402

-3.009

-3.962

-3.134

-3.378

-0.257

-0.035

0.571

-1.551

-1.673

0.051

-5.817

1.149

1.103

2.656

0.177

0.342

-0.007

B Unstandardised

Page 31: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

National Trainee Survey 2006 – key findings 31

Predictors that are statistically significant at P < 0.05 are shaded. The direction ofrelationship is shown by whether a coefficient is positive or negative. All specialtiesare dummy coded with a 1 indicating the trainee is in that specialty and 0indicating they are notx.

As ethnicity is coded 1 for minority ethnic groups and the regression coefficient isnegative, trainees from ethnic minorities are less likely to be satisfied with theirtraining posts. This relationship was tested after accounting for whether the traineequalified outside of the UK (where qualified in Table 8).

With the exception of the Workload Score the indicator scores relate to the OverallSatisfaction Score (see Table 8). This shows that as one would hope and expect theyare measuring facets of the trainees’ experience that relates to their satisfactionwith the post. The finding that the Workload Score is not associated with the OverallSatisfaction Score, suggests that trainees do not object to working hard.

Based on the unique contribution made by the given score (sr2 in Table 8), the mostimportant predictor12 is the Supervision Score, which accounts for 6.7 per cent of thevariance in the Overall Satisfaction Score after accounting for the other measures inTable 8. This finding reflects the importance of the clinical supervisors in ensuringthat the trainees have a positive training experience. Trainees who perceive theirsupervision to be good are more likely to report being satisfied with their post. Ofcourse, if the Adequate Experience Score is included, it is by far the most importantpredictor of the Overall Satisfaction Score (sr2 = 0.264, with a R2 = 0.660, withSupervision Score the second most important. The Adequate Experience Score isonly excluded from this analysis due to measurement issues noted above.

When the scores that were excluded due to the number of cases with missing dataare included in the model, together with the Adequate Experience Score (reducingthe N to 10,416), the pattern remains the same: Adequate Experience is the mostimportant followed by Supervision. The Handover Score was not related to OverallSatisfaction Score; while Career Advice, Feedback, and European Working TimeScores were all related to the Overall Satisfaction Score.

Medical Errors

The Supervision Score is also the most important predictor of a trainee reportingthat they perceived that they had made a serious or potentially serious medicalerror (combined into the Medical Error Score – see Table 5). Logistic regressionwas used to analyse which of the survey’s indicator scores predicted the MedicalError Score (a binary outcome variables scored 0 or 1, see Table 5), the analysisstatistically controlled for the same method and demographic variables as themultiple regression used to predict the Overall Satisfaction Score. Adding theindicator scores into the model in step 3 in Table 9 significantly improves theprediction.

12 Although Adequate Experience was removed

Page 32: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

32 National Trainee Survey 2006 – key findings

Table 9 Overall model predicting the trainees’ reporting of medical errors

Table 10 Survey variables and reporting making a medical error

13 Ethnicity was recoded into white or non-white for the purposes of this analysis.14 Derived from A5 and A6, e.g. respondent working in medicine and intending to practise as consultant in surgery.

Demographicvariables

Time in current postSurvey route

29.33 <0.001 29.33 <0.001

465.850 <0.001 495.18 <0.001

334.790

<0.001

829.97 <0.001

Method variance1

2

3 Survey Indicatorscores

N = 21,252. All specialties included

GradeSpecialty group (dummy coded)Ethnicity13

SexDifferent group from intended14

ForcesYear qualifiedWhere qualified

Access to Educational Resources Adequate ExperienceBullying by ConsultantsEducational SupervisionHours of Education Induction Other Learning Opportunities Supervision Workload

xx2-stepStep Type of variables Variables P xx2 -model P

0.491

3.623

167.984

6.074

0.108

0.468

7.310

1.915

25.062

18.919

5.182

14.775

4.792

0.001

0.174

3.998

1.250

125.957

67.402

7.239

0.483

0.057

0.000

0.014

0.742

0.494

0.007

0.166

0.000

0.000

0.023

0.000

0.029

0.981

0.676

0.046

0.264

0.000

0.000

0.007

1.039

0.938

1.187

0.978

0.846

1.022

0.932

0.748

0.747

0.997

1.007

1.256

0.999

1.007

0.938

1.002

0.978

0.988

0.000

Survey route

Time in post

Specialty group

Different group from intended

Grade (SpR/GPR = 1)

Forces (Forces =1)

Year qualified

Sex (female 1)

Ethnicity (minority ethnic group = 1)

Where qualified (outside UK = 1)

Access to Educational Resources Score

Adequate Experience Score

Bullying by Consultants Score

Educational Supervision Score

Hours of Education Score

Induction Score

Other Learning Opportunities Score

Supervision Score

Workload Score

Constant

Variables that are statistically significant at P < 0.05 are shaded.

WaldPredictor

0.038

-0.064

0.171

-0.022

-0.167

0.022

-0.070

-0.291

-0.292

-0.003

0.007

0.228

-0.001

0.007

-0.063

0.002

-0.022

-0.012

-43.991

B Sig. Odds Ratio

Page 33: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

National Trainee Survey 2006 – key findings 33

In interpreting these findings, it should be noted that the variable being predictedis the trainees’ reporting of perceived medical errors in the last month on thissurvey instrument (and not necessarily their reporting of any incident through theappropriate local procedures for reporting incidents), i.e. not actual medical errors.It is likely that the actual rate of perceived medical errors is higher than reportedon the survey forms, given the method variance finding (see Table 6, Chapter 2).

Predictors that are statistically significant at P <0.05 are shaded on Table 10. Thedirection of relationship is shown by whether a coefficient is positive or negative.There are a number of findings of note. Trainees from a minority ethnic group andtrainees who qualified outside the UK are less likely to indicate they made amedical error in the last month. Trainees who report being bullied are more likelyto report making an error.

It is likely that the effect for specialty group is related to the type of work. Forinstance, as one might expect, emergency medicine trainees report having mademore errors (16.8 per cent, N = 1,217) than trainees in other specialties (9.3 percent, N = 22,829).

Higher Induction, Supervision and Workload Scores are associated with traineesbeing less likely to report making medical errors. This means that a clearer andmore complete induction, better clinical supervision and not being overloaded withwork15 are associated with trainees being less likely to report making medicalerrors. A higher Adequate Experience Score is associated with trainees more likelyto report errors: a job offering plenty of experience is related to trainees beinglikely to report making errors.

Given the importance of the Supervision Score and the Workload Score inpredicting the reporting of medical errors, the individual items were examined;some of these are presented below in Tables 11 and 12. Trainees who reportmaking medical errors are more likely to report saying they are forced to cope withproblems beyond their competence and that they feel that they were supervised bysomeone not competent to do so. They are also more likely to report workingbeyond their rostered hours and feeling sleep deprived.

Table 11 Medical errors and Supervision Score items

15 All Scores were calculated so that a high score was good

10.1 7.1 1.3 21,937No medical errors reported

One or more serious or potentiallyserious medical errors reported

No medical errors reported

How often, if ever, have you beensupervised by someone who youfeel isn’t competent to do so?

Monthly%

How often have you felt forced tocope with problems beyond yourcompetence or experience?

Weekly%

Daily%

60.2

Rarely%

21.4

16.2 14.2 3.1 2,33955.211.2

3.9 2.4 0.8 21,90336.656.4

6.8 4.7 1.0 2,33843.643.9

Never%

N

One or more serious or potentiallyserious medical errors reported

Page 34: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

34 National Trainee Survey 2006 – key findings

PMETB in conjunction with members of the Survey Working Group will publish amore detailed paper analysing these data.

Table 12 Medical Errors and Workload items

The Supervision Score has been shown out of all the indicator scores to be the moststrongly related to Overall Satisfaction Score, Medical Errors Score and ConsultantBullying Score. Like the Overall Satisfaction Score, it varies by grade and specialty.

Supervision Score

To understand what the differences in the Supervision Score mean, one can look atsome of the individual items included in its derivation: 9.7 per cent (184/1,902) ofsurgical SHOs, who as a group are the least satisfied with their supervision,indicated that they were supervised by someone they felt was not competent to doso, monthly or more frequently, whereas only 3.0 per cent (23/772) of the psychiatrySpRs felt that this was the case. Of the surgical SHOs, 26.6 per cent (507/1,909)reported feeling forced to cope with problems beyond their competence orexperience monthly or more frequently, compared with 8.4 per cent (65/775) of thepsychiatry SpRs.

Chart 2 Supervision Score by specialty/grade groups

12.7 31.6 13.5 21,904No medical errors reported

One or more serious or potentiallyserious medical errors reported

No medical errors reported

How often, if ever, have you beensupervised by someone who youfeel isn’t competent to do so?

Monthly%

How often do you work beyondyour rostered hours?

Weekly%

Daily %

33.8

Rarely%

8.4

13.4 37.9 21.3 2,33822.25.1

18.9 15.0 3.4 21,83941.121.6

23.7 23.6 7.1 2,33233.612.0

Never%

N

One or more serious or potentiallyserious medical errors reported

100

Patholo

gy SHO (N =

96)

Psych

iatry

SpR (N =

772)

Patholo

gy SpR (N =

459)

GP (N =

1,851)

Anaesthetic

s SHO (N

= 1,1

85)

Psych

iatry

SHO (N =

1,663)

Paediatri

cs and C

hild H

ealth (N

= 110)

Paediatri

cs and C

hild H

ealth (N

= 1,2

11)

Anaesthetic

s SpR (N

= 1,4

15)

Emerg

ency M

edicin

e SpR (N =

349)

Obstetri

cs and G

ynaecolo

gy SpR (N =

882)

Opthalm

ology SpR (N

= 267)

Surgery

SpR (N =

2,102)

Medic

ine SpR (N

= 2,9

82)

Radiolo

gy (N =

818)

Emerg

ency M

edicin

e SHO (N =

900)

Opthalm

ology SHO (N

= 194)

Medic

ine SHO (N

= 2,9

02)

Obstetri

cs and G

ynaecolo

gy SHO (N =

907)

Surgery

SHO (N =

1,904)

Supervision

Sup

erv

isio

n

Sco

re A

dju

ste

d

95

90

85

80

75

70

65

60

Page 35: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

National Trainee Survey 2006 – key findings 35

Feeling bullied – overall prevalence

To ascertain whether trainees felt bullied, the questionnaire included an itempreviously usedxiii, “J1. Have you been subjected to persistent behaviour in this postthat undermined your professional confidence and/or self-esteem?”

Overall, 10.5 per cent of trainees (N = 23,198) reported being subjected topersistent behaviour in their current post that undermined their professionalconfidence and/or self-esteem. A further 1,460 trainees did not wish to answer andanother 222 left the item blank16, suggesting that the rate may be as high as 16.5 percent, were these responses taken as yes.

The rate is lower than that reported by Paice et alxiv who also used the Hicks item.On their survey, 18.1 per cent of hospital based trainees reported experiencingbullying in their current post, compared with 10.9 per cent of hospital basedtrainees on this survey (N = 20,947, GPRs excluded).

The source of the perceived bullying

Respondents who indicated they had been subjected to this type of behaviour wereasked to indicate its main source. Among hospital trainees (including VTS traineeson hospital placements) this varied by grade. SpR trainees were more likely toreport that a consultant was the source, while SHO grade trainees were more likelyto report that other trainees or nursing staff were the source (see Chart 3):x2 = 307.4, P < 0.001.

16 This was possible on the paper route.

Chart 3 The source of the perceived bullying

1 SHO (Spec and VTS) N = 1,091

2 Spr N = 1,029

The main source of behaviour

Source

70%

60%

50%

40%

30%

20%

10%

0%Consultants Other

traineesNurses or midwives

Managers Patients or relatives

The wholeculture at work

Other

Page 36: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

36 National Trainee Survey 2006 – key findings

Reporting of bullying and subsequent action

Of the trainees who reported bullying, 35.1 per cent (N = 2,268 – missing data in167 cases) indicated that they reported it to their employer/someone in authority; ofthese, 41.7 per cent (N = 575 – missing data in 221 cases) indicated that successfulaction was taken to stop it. Of all the reported incidents (all incidents includingthose for which there are missing responses to the reporting and successful actionitems) only 9.9 per cent (N = 2,435) were reported and subjected to remedialaction.

Whether a trainee who experienced bullying reported it varied according to thesource of the bullying. Trainees were more likely to report it if it came frommanagers or patients/relatives than if it came from other clinical staff (x2 = 21.41, P < 0.001, N = 1,757 cases of bullying with a specific source only). 34.9per cent of bullying involving trainees, consultant or nursing staff was reported,compared with 54.0 per cent of bullying involving managers or patients/relatives.

Have you been subjected to persistent behaviour in this post that has underminedyour professional confidence and/or self-esteem?

20%

18%

16%

14%

12%

10%

8%

6%

4%

2%

0%

Obstetri

cs and G

ynaecolo

gy SpR (N =

793)

Obstetri

cs and G

ynaecolo

gy SHO (N =

844)

Emerg

ency M

edicin

e SHO (N =

814)

Surgery

SHO (N =

1,745)

Anaesthetic

s SHO (N

= 1,0

92)

Anaesthetic

s Spr (

N = 1,3

43)

Psych

iatry

SHO (N =

1,583)

Radiolo

gy SpR (N =

765)

Patholo

gy SHO (N =

765)

Psych

iatry

SpR N (N

= 747)

Opthalm

ology SHO (N

= 178)

GP (N =

1,782)

Patholo

gy SpR (N =

428)

Emerg

ency M

edicin

e SpR (N =

335)

Medic

ine SHO (N

= 2,7

10)

Medic

ine SpR (N

= 2,8

19)

Paediatri

cs and C

hild H

ealth SHO (N

= 1,1

29)

Paediatri

cs and C

hild H

ealth SpR (N

= 1,0

26)

Surgery

SpR (N =

1,983)

Opthalm

ology SpR (N

= 245)

Chart 4 Bullying by specialty/grade group

Are certain types of trainee more likely to be bullied?

Trainees’ reporting of bullying also varies by specialty and grade, as illustrated inChart 4 below (xx2 = 175.37, P < 0.001. N = 22,450)17. The obstetrics and gynaecologytrainees report the most bullying, while GPs report the least.

Confidence interval calculated following Altman et alxv (2000)

These findings concur with existing literature that has shown an increasedprevalence of “harassing or discriminatory behaviours” in surgery and obstetricsand gynaecologyxvi.

Logistic regression was used to look at the relationships between feeling bulliedand demographic variables. These relationships varied by specialty group (seeTable 13).

17 For clarity on the chart, very small or non-existent groups were excluded: public health, occupational medicine,and radiology SHOs.

Page 37: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

Table 13 The relationship between demographic variables and reportingbullying for each specialty group and overall

Table 13 gives the odds ratios for each variable that is related to trainee reportingbullying at P < 0.05.Where no odds ratio is given, the relationship is not statisticallysignificant. For instance, for surgical trainees the following variables are associatedwith an increased probability of reporting bullying on this survey: being in postlonger, qualifying outside the UK and being female. None of the other variables areassociated with reporting bullying.

With the exception of time in post and year qualified, all variables are binary,where they are coded 1 for the presence of the description given in Table 13 (e.g.female = 1). The Odds ratio for time in post and year qualified indicate that thelonger the trainee has been in post, the more likely they are to report being bulliedin that post and trainees who qualified more recently (i.e. left medical schoolrecently) are more likely to report being bullied.

For binary variables, the odds ratios show the increased probability of reportingbullying if the factor is present. For all specialties, trainees from a minority ethnicgroup are 1.3 times more likely to report being bullied than white trainees.

In logistic regression, when all the variables are entered at the same stage, eachvariable is tested after accounting for all the other variables in the model, soqualifying outside the UK and being a non-white trainee increases the probability ofbeing bullied. This is illustrated in Table 14.

National Trainee Survey 2006 – key findings 37

Anaesthetics

Traineeis locum

Traineein HMforces

Yearqualified

Timein post

2,284

N QualifiedoutsideUK

1.51

Emergency medicine(model not significantat P <0.05)

1,099 1.56

Female Traineenot white

4.11General Practice 1,754 1.82

Medicine 1.235,287 1.34

1.99 1.04Obstetrics andgynaecology 1,535 1.71

11.59Ophthalmology 401

Paediatrics andchild health

1.412,063

Pathology (model notsignificant at P <0.05) 477

1.59Psychiatry (model notsignificant at P <0.05) 2,176

Radiology 1.98744 1.92

1.33Surgery 1.203,489 1.56

1.01 1.15All specialtiespredicting anybullying.

1.2121,758 1.11 1.30

1.42 0.95 0.74All specialties -predicting consultantbullying

1.5021,758

Variables in model. If the variable is related (P <0.05) to reporting bullying theOdds ratio is reported, if it is not related the cell is blank.

Page 38: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

38 National Trainee Survey 2006 – key findings

Looking at reports of feeling bullied by consultants only, in the final row of Table 13,gives a different pattern; neither the trainees’ sex nor their ethnicity (white versusnon-white) is associated with the trainees reporting bullying by a consultant.Furthermore, trainees from outside the UK are less likely to feel bullied byconsultants (odds ratio = 0.74) and older trainees are more likely to report feelingbullied by consultants (this fits with the finding that SpR are more likely to feelbullied by consultants – see Chart 3).

This pattern of results reflects trainees’ own perceptions. Of those who reportedbeing bullied and answered whether this was related to their ethnicity, sex, sexualorientation or religious beliefs, 36.1 per cent of trainees who felt bullied by anynon-consultant source felt it was related to one of these, against 28.3 per cent ofthose who reported the source as being a consultant (N = 1,888, x2 = 12.8,P < 0.001).

The model varies across specialties; the trainee’s gender is only related toperceived bullying within certain specialty groups - anaesthetics, emergencymedicine and surgery. GP trainees working in the armed forces are more likely toreport perceived bullying than civilian GPs (seven of the 22 GPs in the armedforces answered yes to item J1).

Perceived bullying and indicator scores

Respondents who reported being bullied had lower scores on all the indicatorscores, except the Handover Score (all mean comparisons were statisticallysignificant at P 0.05 after the Bonferroni correction was applied), suggesting thatexperiencing bullying is associated with a generally less positive training post. Thelargest association was with the Supervision Score, again indicating the importanceof this indicator.

PMETB in conjunction with members of the Survey Working Group will publish amore detailed paper analysing these data. It is worth noting here that, as Paice et alnote, some of the behaviours that erode trainees’ professional confidence or self-esteem may be attempts to improve trainees’ performance, so an educational ratherthan punitive approach is required to tackle the problem.

Table 14 Trainees’ ethnicity and place of qualification and bullying

9,966White qualifying within the UK

White qualifying outside the UK

Minority ethnic group qualifying within the UK

Minority ethnic group qualifying outside the UK

% reporting bullying

8.6

N

1,5699.6

3,40910.8

7,19511.9

Page 39: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

National training initiatives

The survey data were used to assess the impact of two national initiatives that havesought to improve the training experience of junior doctors:

● Radiology academiesxvii

● Hospital at Nightxviii

In both cases, some of the predicted differences emerged, providing furtherevidence of the construct validity of the scores derived from the survey data andhighlighting their potential for evaluating other initiatives that are designed toimprove junior doctors’ training posts.

Radiology academies

Three academies were delivered as part of the ‘Radiology - Integrated TrainingInitiative (R-ITI)’, a national programme to provide an increased number of highquality radiologists.

Numbers of respondents

Only trainees in clinical radiology (also known as diagnostic radiology andformerly known as radiology) were included. The four clinical radiology traineeswhose grade was recorded as something other than SpR were excluded. This left 68trainees in the academies (as detailed in Table 15) and 536 trainees not in theacademies. Trainees at the academy sites but not in the 2005 or 2006 cohorts oftrainees have also been classified as training at an academy.

Approach to analysis and results

The mean indicator scores (derived from survey items – see Table 5 in Chapter 2)for the two groups were compared and the ANOVA (F-test) used to test fordifferences; the Bonferroni correction for multiple testing was applied. A number ofthe mean scores were significantly different at P < 0.05; they are shaded in Table 16below. All the differences are in the expected direction. Trainees based within anacademy reported a more positive training experience. The survey was notspecifically designed to evaluate the radiology academies and it is likely that asurvey with items concerned specifically with the training of radiologists wouldhave found more marked differences.

National Trainee Survey 2006 – key findings 39

Table 15 Respondents by radiology academy

24Leeds and West YorkshireRadiology Academy

HospitalsAcademy

Leeds Teaching Hospitals NHS Trust

16Norfolk & NorwichRadiology Academy

Norfolk & Norwich University Hospital NHSTrust

N

2215

Peninsula RadiologyAcademy

Plymouth Hospitals NHS TrustRoyal Devon & Exeter NHS Foundation TrustSouth Devon Healthcare NHS Trust

Page 40: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

40 National Trainee Survey 2006 – key findings

Ta

ble

16

Com

pa

riso

n b

etw

een

aca

dem

y a

nd

non

-aca

dem

y t

rain

ees

on t

he

Na

tion

al T

rain

ee S

urv

ey I

nd

ica

tors

1.00

1.00

1.00

1.00

1.00

1.00

1.00

1.00

0.00

1.00

1.00

0.04

0.03

1.00

0.02

Acc

ess

to

Ed

ucat

ion

al R

eso

urce

s Sc

ore

Ad

eq

uate

Exp

eri

en

ce S

core

Sco

re

Car

ee

r A

dvi

ce S

core

Bul

lyin

g %

re

po

rtin

g

Ed

ucat

ion

Sup

erv

isio

n S

core

Eur

op

ean

Wo

rkin

g T

ime

Dir

ect

ive

% "

Yes"

Fee

db

ack

Sco

re (

Ad

just

ed

)

Han

dov

er

Sco

re (

Ad

just

ed

)

Ho

ur E

duc

atio

n S

core

Ind

ucti

on

Sco

re

Oth

er

Lear

nin

g O

pp

ort

unit

ies

Sco

re (

Ad

just

ed

)

Ove

rall

Sat

isfa

ctio

n S

core

(A

dju

ste

d)

Sup

erv

isio

n S

core

(A

dju

ste

d)

Wo

rk I

nte

nsi

ty S

core

(A

dju

ste

d)

Wo

rklo

ad S

core

(A

dju

ste

d)

me

an s

core

at

loca

tio

n

P (

wit

hB

onfe

rron

i)

2.90

1.47

0.50

0.28

0.04

0.02

0.76

1.11

30.6

6

0.21

0.01

9.14

9.76

0.11

10.9

2

F

12.3

9

11.7

5

33.7

8

17%

1.11

37%

20.6

1

14.9

3

2.70

0.80

16.3

9

9.90

12.2

0

22.0

9

17.5

2

Sta

n D

ev:

Aca

dem

y

68 68 58 68 68 56 57 23 68 68 67 68 68 68 68N:

Aca

dem

y

83.0

2

74.9

0

85.3

4

3% 3.96

84%

53.7

8

33.0

1

4.75

2.51

51.6

6

82.3

5

88.9

9

88.7

8

64.7

8

Mea

n:

Aca

dem

y

14.1

8

13.0

1

36.4

3

20%

1.15

38%

23.7

6

18.4

1

1.87

0.76

15.2

1

12.6

7

13.5

8

19.4

6

14.4

4

Sta

n D

ev:

non

Aca

dem

y

536

536

403

536

535

444

436

276

535

534

521

536

536

535

534

N:n

onA

cad

emy

79.9

5

76.9

1

81.7

6

4% 3.99

83%

50.9

0

28.8

6

3.34

2.56

51.4

3

77.5

2

83.5

8

89.6

4

58.4

7

Mea

n:

non

Aca

dem

y

Sig

nif

ican

t d

iffe

ren

ces

in m

ean

sco

res,

P <

0.0

5 ar

e s

had

ed

ab

ove.

Page 41: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

The Hospital at Night

The Hospital at Night initiative aims to reduce dependency on training gradedoctors for providing cover at night, in order to reduce their working hours andensure that these are compliant with the European Working Time Directive18, whileensuring that there is no negative impact on their training. Hospital at Nightadvocates supervised multi-disciplinary handover in the eveningsxix.

The impact of the Hospital at Night programme was assessed by comparingtrainees working in hospitals with Hospital at Night teams with trainees in otheracute hospitals that do not have these teams. Locations were classified asparticipating in Hospital at Night on the basis of data supplied by the Hospital atNight team19. On the basis of Hospital at Night’s stated objectives, particularindicator scores and items were compared across these two groups of trainees.

National Trainee Survey 2006 – key findings 41

18 http://www.dh.gov.uk/en/Policyandguidance/Humanresourcesandtraining/Workingdifferently/Europeanworkingtimedirective/index.htm.

19 Data kindly supplied by Gerry Bolger, Project Director - National Hospital at Night Team

Table 17 Handover

Hospital at Night respondents should report having multi-disciplinary handovers.These data suggest that this is the case (Table 17 below): 23.7 per cent reportedthat nurses are involved in handovers at Hospital at Night locations compared with11.7 per cent at non-Hospital at Night providers. The difference is more marked ifthe comparison is restricted to medical trainees.

All traineesat acutesites

Medical trainees at acute sites

Surgical trainees at acute sites

P(Bonferroniapplied)

xx2Anorganisedmeeting ofdoctors andnurses %

Anorganisedmeeting ofdoctors %

A phone oremailcommunication%

Informal%

None %

Not implemented N = 11,388

ImplementedN = 4,028

Not implementedN = 3,393

ImplementedN = 1,228

Not implemented N = 2,266

ImplementedN = 725

1.9

1.9

1.2

2.0

1.6

2.2

38.5

28.0

44.4

20.0

35.5

30.1

4.2

3.8

5.2

5.5

5.3

5.5

43.7

42.7

32.0

24.3

49.9

49.8

11.7

23.6

17.1

48.2

7.8

12.4

378.63

507.07

19.33

0.00

0.00

0.00

Implementation status

Which of the following best describes handover arrangements BEFOREnight duty in your post?

Page 42: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

42 National Trainee Survey 2006 – key findings

Table 18 Impact on working times

The Hospital at Night programme aims to help trusts comply with the EuropeanWorking Time Directive. The data below do not suggest this, as the difference incompliance is in the opposite direction to that predicted and the differences,although small, are statistically significant for all trainees at acute sites and the sub-group of medical trainees at acute sites. Similarly, there are small statisticallysignificant differences on the second item for all trainees and the sub-group ofmedical trainees; trainees working at Hospital at Night sites are slightly more likelyto have been asked to submit hours that are compliant, when the hours actuallyworked were not compliant..

Obviously this finding is only an association. It may be the case that Hospital atNight providers knew there was a problem in complying with the EuropeanWorking Time Directive and, as a result, have started to tackle it, but the changeshave not yet bedded down sufficiently to impact upon working times. The analysiscan be repeated with the 2007 data to test this hypothesis and would bestrengthened by more detailed data on the implementations of the initiative atHospital at Night providers

The difference between night to day handover between trainees at Hospital at Nightsites compared with trainees at other acute sites was far less marked. The biggestdifference was for the medical trainees; for this group Hospital at Night sites had2.7 per cent more respondents saying their handover involved nurses. As the focusof Hospital at Night is on the day to night handover, these data would seem toreflect changes to handover process related to Hospital at Night implementation.

All trainees at acute sites

Medical trainees at acute sites

Surgical trainees at acute sites

P (Bonferroni)xx2N% YesP (Bonferroni )xx2N

Not implemented

Implemented

Not implemented

Implemented

Not implemented

Implemented

10,834

3,750

3,086

1,132

2,150

668

% Yes

89.0

87.4

88.4

84.9

82.0

81.4

7.160

9.145

0.090

0.045

0.015

1.000

18.9

21.8

19.9

23.6

24.5

24.8

11,716

4,010

3,371

1,210

2,371

725

16.064

7.080

0.023

0.000

0.047

1.000

Implementation status

Are your rostered working hourscompliant with the European WorkingTime Directive?

Have you been asked to submit hoursthat are compliant with the EuropeanWorking Time Directive, when thehours you actually work are NOTcompliant?

Page 43: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

National Trainee Survey 2006 – key findings 43

Table 19 Work intensity

The work intensity items reflect a similar pattern, with trainees at Hospital at Nightsites reporting a heavier workload during the night. Again, the differences are smallbut, due to the sample sizes available, they are statistically significant. Thedifferences in workload during the day are less reliable; the difference is onlystatistically significant when all trainees are included in the analysis.

Table 20 Impact on training

The Hospital at Night initiative is not meant to impact on doctors’ training and thesurvey data suggest that, on balance, it does not affect trainees’ perceptions of theexperience they are getting, as there are no consistent statistically significantdifferences between trainees at sites that have or have not implemented Hospital atNight on the Adequate Experience Score (derivation of scores is outlined in Table5, Chapter 2).

All trainees at acute sites

Medical trainees at acute sites

Surgical trainees at acute sites

P (Bonferroniapplied)xx2N

%respondingheavy orvery heavy

P (Bonferroniapplied)xx2N

Not implemented

Implemented

Not implemented

Implemented

Not implemented

Implemented

13,290

4,651

3,829

1,380

2,700

853

% respondingheavy or very heavy

Day Night

37.6

41.9

41.0

44.8

34.4

34.1

28.08

7.99

2.70

0.00

0.55

1.00

39.1

44.8

47.0

52.7

25.1

31.2

12,132

4,163

3,460

1,219

2,597

804

49.05

15.65

15.37

0.00

0.02

0.02

How would you rate the intensity of your work by:

All traineesat acute sites

Medicaltrainees atacute sites

Surgicaltrainees atacute sites

P (Bonferroni applied)FNSD

Not implemented

Implemented

Not implemented

Implemented

Not implemented

Implemented

16.06

16.63

15.32

16.27

19.30

19.90

Mean

73.54

72.99

72.58

72.73

70.35

68.49

13,330

4,657

3,842

1,384

2,702

853

4.02

0.09

5.95

0.14

1

0.04

Adequate Experience Score

Chi-squared test run across all five response options. Presented as % heavy/very heavy in this table for ease interpretation only

Page 44: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

44 National Trainee Survey 2006 – key findings

Deaneries, training provider and differences in OverallSatisfaction and Supervision – a multilevel model analysis

These data are inherently multilevel: trainees work within departments, which arewithin hospitals that fall within a given deanery’s responsibility. PMETBconsequently commissioned multilevel modellingx to analyse these data at differentlevelsxi. This analysis looked for differences between deaneries, training providersand specialty groups within training providers on two of the indicator scores:Overall Satisfaction and Supervision.

The model included a number of background variables to statistically control fortheir effects, so that differences between providers cannot be attributed todifferences between them on these background variables.

The model has four nested hierarchical levels that reflect the organisation ofpostgraduate medical education: trainees; within specialty groups within providers(e.g. Medicine at a given provider); within providers; within deaneries. Significantvariation was found on the Overall Satisfaction Score and the Supervision Score atthe level of specialties within providers (P < 0.001 and P < 0.001) and providers (P= 0.001 and P = 0.005). There is no difference on either measure at the level ofdeaneries. The differences do not reflect differences in trainee mix in terms of thebackground variables: specialty, sex, grade, type in post, years qualified, or route ofresponding to the questionnaire.

While these results show that there are differences in trainees’ perceptions of thequality of their training associated with training providers and specialties withintraining providers, it is not possible to use these data alone to reliably identifypoorly performing training providers. This suggests that, while the survey data mayhave utility as a screening tool as proposed by PMETB this year, it would beinappropriate to take action or identify poorly performing training providers on thebasis of these data alone.

Page 45: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

National Trainee Survey 2006 – key findings 45

4. The survey data as a quality management tool

As the primary objective of the survey work has been quality improvement, PMETBhas focused on releasing data to deaneries, which are responsible for local qualitymanagement. Deaneries have been working with training providers within theirarea on action plans in response to the survey’s findings. Table 21 details the datareleased to deaneries that was used for this work.

Data releases

Table 21 Data that PMETB has released

Data Release data and format Audience

Data by specialty group andprovider

PMETB/COPMeD NationalScreening Tool with indicatorscores for each specialty groupfor all providers in the UK withmore than three respondentsderived from the survey data.

Specialty group data byprovider benchmarked tonational means and quartiles

The indicator scores (used on theCompare CD above) withcomparison with the nationalmeans and quartiles for the givenspecialty group; these means andquartiles use the data from allrespondents, including thosewhere there are less than threerespondents.

5 December 2006

Compare CD –

Compare software allows the user toexplore these data and producereports as Word documents for agiven location and specialty group.The user can choose which otherlocations/VTSs to compare theirchosen location/VTS with on thecharts.

7 December 2006

Screening Tool Access Databasethat allows the user to obtain areport for a group of indicatorswhere scores for all availablespecialty groups are displayedfor each location.

Deaneries

Deaneries

Specialty data

Indicator scores aggregated tospecialty and grade groups bydeanery.

This dataset contains indicatorscores for each specialty andgrade group in each deanery,where there are more than threerespondents available, e.g.scores for cardiology SpRs inLondon Deanery. This report willtherefore include specialtiessuch as public health andoccupational medicine that, dueto low numbers at individualproviders, have not beenincluded in data by specialtygroup and location.

11 January 2007

Compare CD

Royal Colleges

Page 46: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

46 National Trainee Survey 2006 – key findings

Data Release data and format Audience

Specialty data – UK wide

Indicator scores aggregated tospecialty and grade for thewhole of the UK.

This dataset contains indicatorscores for each specialty andgrade group across the UK,where there are more than threerespondents available, e.g.scores for cardiology SpRs UKwide.

Provider profile report

A summary of each provideracross all their specialty groups.It identifies the providers withthe highest proportion of outlierindicator scores (both aboveand below the mean) across allthe specialty groups.

11 January 2007

Compare CD

1 February

Excel

Royal Colleges

Deaneries

Deanery level analysis oftraining since leavingmedical school

An analysis based on responseto item G4 by deanery.

21 December 2006

Excel

Deaneries

NHS Institute forInnovation andImprovement

Intended specialty analysis 18 December 2006

Excel spreadsheet available here:http://www.pmetb.org.uk/index.php?id=intendedspecialtyanalysis

Trainees

Deaneries have been using the data released to them for their own qualitymanagement activities. PMETB has not been prescriptive about how deaneriesshould go about this and generally deaneries have discussed and disseminated thedata locally as they deem appropriate. It is important for deaneries to undertakethis activity, as a national organisation such as PMETB would not be able to interpretthe data within the context of the NHS locally.When considering the reason for alow score on an Indicator Score derived from survey data, it is advisable to excludecauses such as particular local circumstances (for instance a change to theconfiguration of services that has been disruptive) before suggesting the score isindicative of a performance issue. Only deaneries would be able to take suchcauses into consideration when planning follow up on the survey data.

Page 47: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

National Trainee Survey 2006 – key findings 47

MerseyEasternWest MidlandsNorth WesternOxfordYorkshireEasternScotland (West)Scotland (West)OxfordLondonScotland (West)

ClassificationUpper 95%Lower 95%N

123456789101112

192312121239607101675

MeanDeanery Provider

69.9067.3067.0069.0671.6574.4474.5068.3371.0570.8570.6266.01

60.2859.5356.0661.1665.7869.5569.9655.7857.3561.1960.2251.33

79.5375.0777.9376.9777.5379.3279.0480.8884.7680.5281.0280.70

Outlier

Below mean and 2nd

quartile Not outlier

Bottom quartile Not below mean Not outlier

Table 22 Surgical providers in the bottom quartile and/or below the nationalmean based on the confidence intervals

20 Adjusted for method variance – see Chapter 2

The identification of outliers for follow up

PMETB only intends the survey data to be used as an initial screening tool, toidentify areas that may require further investigation. For the purposes of identifyingoutliers, the following analysis was conducted.

The survey data were aggregated to the level of specialty groups within providers,so for example there were mean adjusted20 scale scores for surgical trainees at agiven provider. To ensure any given trainee’s responses remained anonymous aspromised, scores were only calculated where there were more than threerespondents working in the specialty group at the given provider. This is datarelease, data by specialty group and provider, in Table 21.

The mean scores for each provider’s specialty group of trainees were comparedwith the national mean and quartiles for that specialty group (as per specialtygroup data by provider benchmarked to national means and quartiles in Table 21).The national comparison group included all trainees within that specialty group,including those working at providers with less than three respondents. So, forinstance, the mean score for the given provider’s surgical trainees was comparedwith the national mean and quartiles for all surgical trainee respondents across theUK. A score was defined as outlying if it was in the bottom quartile and below thenational mean (based on the confidence intervals not overlapping). The advantageof including data from the confidence intervals in the comparison is that, for asample from a particular provider, the confidences are narrower if there are morerespondents and if there is more agreement among those who have responded. Justusing the quartile information takes no account of sampling error and just using acomparison of the means meant that it was possible for providers not in the lowestquartile to be deemed outliers because they had very narrow confidence intervals.Therefore using both pieces of information was felt to be more appropriate.

An example of identifying outliers – surgical specialty group

For the all the surgical group trainees in the UK (N = 4,022), the mean SupervisionScore is 80.58 (95 per cent confidence intervals 80.11 to 81.06). The 25th percentileis 71.09. The 209 providers for whom there were four or more respondents werethen compared with these figures. Table 22 shows which providers were classifiedas outliers because they met both criteria, and the providers that only met one ofthe two criteria and were therefore not classified as outliers. Providers 1 to 4 areoutliers; providers 5 to 13 are not.

Page 48: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

48 National Trainee Survey 2006 – key findings

The proportion of outlying indicators across all available specialties was calculatedfor each provider. Of the providers in this analysis, 34 per cent (120/355) had noindicator scores that were outliers (taken from the Provider profile report). Only 15per cent (52/355) had 10 per cent or more of indicator scores across all theirspecialties with respondents that were outliers.

Deanery quality improvement

All deaneries were asked to provide an action plan in response to the deaneryspecific data provided from the National Trainee Survey. The deaneries were givena proforma with headings that are standard in most action plans (see Appendix 2for a blank proforma at www.pmetb.org.uk/traineesurvey).

The deaneries were given three months to return an action plan to PMETB21. Alldeaneries achieved a response, many with a full and considered action plan. InScotland, NHS National Education for Scotland (NES) provided a Scotland wideresponse and then, in two cases, a deanery specific overview. The responses werevariable in style and content. This was expected at this early stage and with thechallenges of effective dissemination from PMETB to the deaneries, and from thedeaneries to the education providers (e.g. NHS trusts and boards).

The data received by deaneries were affected by the filter to keep trainees’identities anonymous, in that specialty groups at providers for which less than fourresponses were received were not included in the report. For some specialtygroups at some providers, deaneries have noted that more information is requiredto link the trainees’ responses to a specific clinical environment/department andthus write a detailed action plan. PMETB will explore whether some specialties (asopposed to specialty groups) can be reported for some larger providers instead ofjust across the deanery. However, it will not be possible to report on something suchas a department within a provider that is not defined by a provider listed by theNational Administrative Codes Service and a specialty or sub-specialty.

Equally, deaneries need to adapt and plan their responses. Several of the deaneriesdecided to document their actions in order to create a full action plan which theywill send to PMETB at a later point – an ‘action plan for action plans’.

All deaneries used the outlier identification detailed above as the rationale forinclusion of a provider within an action point. The action plans were very clear onthe levels for responsibility and generally the headings do not appear to havecaused problems. Some of the deaneries were notable for their clarity of bothresponsibility and timescales.Virtually all deaneries evidenced a clear grasp ofpriorities and risk management, and there was significant consistency in theidentification of high risk and appropriate timescales. One deanery used a threelevel approach, with urgent action within one month as the first level, the secondlevel for action within three months and the remainder as developmental changes.

Several of the deaneries documented their concerns about disseminating the datain a useful way to their local education providers and PMETB will explore thisfurther when consulting on reporting (see Chapter 5). Half of the deaneries felt ableto be specific in identifying the education providers concerned. Several of thedeaneries referred very helpfully to previous quality management activity and/orthat the data matched their experience of the quality of education for thoseeducation providers. Occasionally, the deaneries have questioned a mismatchbetween their own and the education provider’s experiences and the outcomes inspecific areas such as handover. In addition to this retrospective analysis, severaldeaneries noted future activity which could occur where there was evidence of

21 Proforma sent out 7 December 2006 with specialty group data by provider benchmarked to national means andquartiles (see table 21)

Page 49: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

National Trainee Survey 2006 – key findings 49

continuing problems. One deanery noted that the next step could be initiation of atriggered PMETB visit for one education provider.

The utility of specialty group based data and identification of specialty groupsthemselves was also interesting and will be considered in future development ofthe surveys. Some (less than 50 per cent) of the action plans identified the specialtygroup clearly and specifically. Other action plans did not identify the providerspecialty group clearly, concentrating on the issue and the education provider(s).Several linked specialty group, the issue highlighted by the indicator score and theprovider; this approach makes the action plan both understandable and easier tofollow up. Once action planning is embedded in the deanery systems, theinvolvement of specialties and schools will become the norm. One deanery hadvery clear identification of the role of the schools and Specialist TrainingCommittee chairs within the action plan; this was one of the more successful inintegrating issue, specialty and provider.

Deaneries will use the action plans for several purposes, including deanery qualitymanagement, but also to provide the information and reassurance that PMETB willneed that Generic standards for training are being attained and maintained. Formneeds to follow function; it may be that action plans at a higher level and withbroader scope will meet regulatory and performance management issues, while amore detailed approach is needed at quality management (deanery) and qualitycontrol (provider) levels. A lack of detail in many of the action plans means thatexternal readers cannot assess accuracy, the relevance of issues or appropriatenessof actions taken; neither can they identify if the actions have been undertaken.

Many of the deaneries and NES confirmed their intention to improve both theresponse rate and the accuracy of their population data (an issue discussed inChapter 5).

Page 50: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

50 National Trainee Survey 2006 – key findings

5. The future of the National Trainee Survey

The National Trainee Survey of 2006 was the first national survey conducted byPMETB and COPMeD and there were significant challenges in its delivery and anumber of important lessons learnt. The high response rate and the debate anddiscussion that followed the release of information so far indicates that there is asignificant role for this survey in the development and quality assurance ofpostgraduate medical education in the future.

A number of changes are planned for the 2007 survey; these are described belowand are based on PMETB’s experience of the 2006 survey and feedback from thedeanery staff that kindly helped with the work.

Route of administration

To eliminate the method variance problem outlined in Chapter 2, where trainees’responses to some items varied by the route by which the survey was administered,the 2007 survey will only be administered using one route: a national website.Webadministration of the survey has a number of benefits over the other methods usedfor the 2006 survey. These are:

1. It is more economical than scanning paper forms and the portable electronicsurvey units.

2. It allows the easy administration of specialty specific versions of thequestionnaire as items can be conditionally displayed based on the responses toearlier items.

3. The system can produce a reference number so that doctors can prove they havecompleted the survey, thereby providing deaneries with the data to enforce themandatory status of the future trainee surveys.

4. A web based survey will dramatically reduce local administrative burden. Thetask managed locally will be the collection of valid email addresses (which willhave considerable collateral benefit).

It is known that trainees may alter their responses when they feel these could beidentified (although assurances were given that the data were confidential andreporting would ensure that responses remained anonymous); this might reduce thetotal variance available for analysis from some items. However, this will notinvalidate comparison between training providers, as it may be assumed that anymeasure of it is not associated with any particular provider since all are using thesame method of administration and impression management22 is an individual-levelvariable. There may be inter-specialty differences in impression management but,as benchmarking is within specialty group, any such differences will not bias theanalysis.

The content of the survey

Changes to the content of the survey are constrained by two factors:

1. The data will be used to test for changes over time; therefore the items used toderive the indicator scores must remain the same.

2. Shorter questionnaires are known to elicit a higher response ratexx and thus thequestionnaire should not be substantially lengthened.

Within these constraints a number of changes are proposed for the 2007 work.

22 Impression management: the extent to which an individual answers items to give a more positive impression ofthemselves.

Page 51: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

National Trainee Survey 2006 – key findings 51

Adequate experience

The sine qua non of any post is to provide the trainee with the experience s/heneeds to acquire the competences set out in the curriculum that s/he is following.Currently, this standard is addressed through two summative evaluation items onwhich trainees are asked to rate the quality of the experience overall. It has notbeen possible to demonstrate the construct validity of these items; it cannot beshown using factor analysis that the items capture something different from overallsatisfaction with the post – a failure of discriminant validity. This means that thescore is not adding substantial information to that which is already contained in theOverall Satisfaction Score. Furthermore, it does not provide diagnostic informationon which aspects of the experience offered by the post were found to beinadequate. As all trainees received the same items in 2006, it would not have beenpossible to have items that related to facets of training experience, as these arespecialty specific. However, in 2007, it will be possible to display itemsconditionally, based on the respondent’s earlier answers. It will therefore bepossible to ask about particular facets of experience in relation to both the trainee’sspecialty and stage of training. Item development for these speciality-specificadequate experience items will be done in conjunction with lead deans and theRoyal Colleges and faculties. Items will be related to the PMETB approvedcurricula. The challenge will be to devise items that will provide sufficient datavariability to distinguish between posts providing experience that traineesperceive as ‘good’ and posts that are not perceived as ‘good’. Posts are part of aprogramme that is designed to provide the necessary experience overall, and thusit may be necessary to ask whether the post provided the experience the traineeexpected, given the design of the overall programme.

Overall Satisfaction Score

There is a ceiling effect on the Overall Satisfaction Score, which means it is notpossible to distinguish between posts at the top end of the score. All doctors whoare satisfied with their posts are alike as far as the survey data is concerned, 12.4per cent (N = 24,848) of respondents having the maximum possible score on thisscale. This measurement issue will be addressed in the 2007 survey with theaddition of items written with a view to making distinctions possible at the positiveend of the overall satisfaction score.

Follow up of concerns

The survey currently promises confidentiality for respondents, but this promiseconflicts with some respondents’ expressed desire for action to be taken inresponse to concerns they have raised in free text comments. Therefore, the 2007survey will allow respondents to indicate that they would be happy to be identifiedin the course of any follow-up action in response to their concerns. This approach istaken by the Training and Development Agency for Schools in their newly qualifiedteacher survey, which also examines the perceived quality of training receivedxxi.

Personality variables

From the occupational psychology literature it is known that a respondent’spersonality accounts for some of the variance in their self-report of constructs suchas job satisfactionxxii. Incorporating a personality measure in the model can improvethe prediction of turnoverxxiii; it might therefore be expected that incorporating apersonality measure would improve the trainee survey’s predictive validity (as the

Page 52: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

52 National Trainee Survey 2006 – key findings

effects of the personality measure can be statistically adjusted). Negative ratingsfrom trainees who are generally positive but negative about their current trainingpost are more likely to relate to problems with the posts than negative ratings fromtrainees who are generally negative. The 2007 survey will therefore include ashortened measure of positive and negative affect to allow for statistical adjustmentof personality factors when using the Overall Satisfaction Score to benchmarktraining providersxxiv.

Responses varied by route of administration. For some items, more positiveresponses were more likely with less anonymous routes of survey administrationand thus it would seem appropriate to include some items to measure the extent towhich trainees are considering the impression of themselves that they arepresenting when answering. This measure of impression management can then betested for associations with items used in the indicator scores and, if necessary,statistically controlled for. Merrill et alxxv have developed a scale designed tomeasure impression management that is domain specific for medicine, which couldbe used by the 2007 survey.

Dissemination of the results

The time from data collection to reporting of the 2006 data was longer than it shouldhave been. Reporting was impeded by the following factors, all of which will not bepresent for the 2007 work:

1. Four routes of administration, each with variance from the master data template,had to be merged to create one data set.

2. Indicator scores had to be adjusted for method variance to allow comparisonacross training providers because the provider was confounded by the route ofadministration.

3. Some respondents’ data had to be remapped to an alternative location. Forinstance, in one English Postgraduate Medical Education Centre, GPRs hadreceived paper forms with the acute trust identifier.

It should therefore be possible to report the findings back to the deaneries andother stakeholders in a more timely fashion. A variety of data aggregations areappropriate, given the local systems in place for quality managing postgraduatemedical education; these are described below in Table 23. Each aggregationreports on the same indicator scores. Aggregations are filtered, in that onlyproviders or rotations with more than three respondents are reported to ensure thatindividual trainees’ responses remain anonymous.

Table 23 Aggregations for reporting

Audience Aggregation

Deans and Directors of Medical Education Training provider and specialty group, for example anaesthetics trainees at St Elsewhere NHS Trust

Deans and Specialist Training Committees Deanery Specialty Grade, for example London Cardiology SpR trainees

Page 53: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

National Trainee Survey 2006 – key findings 53

The Deanery Specialty aggregation ensures that the responses from trainees in thesmaller specialties are reported on, as frequently there would not be enoughresponses to include this in the Training Provider and Specialty Group aggregation.For example, with 2006 data, occupational medicine was reported by DeanerySpecialty Grade aggregation only and not by location, as there are not enoughtrainees at a given location to get above the reporting filter.

As the focus of the work is on quality management, reporting back to those withaccountability for the training and seeking their responses to the data shouldalways be the priority. However, other parties, not least the trainees who completedthe survey, have a right to see the data for the locations of interest to them. Tofacilitate this, PMETB plans to report local findings through a web based reportingservice. This will supplement or replace (depending on whether all the requiredfunctionality can be obtained) the existing method which uses a CD-ROM and theCompare software provided by the Healthcare Commission. PMETB will consult onthe reporting options for the 2007 survey as part of the quality assurance (QA)framework consultation to establish:

● that the aggregations are at a level of granularity that allows data to be linkedback to a clinical area for which an action plan can be written; constraints includethe identification of the entity to aggregate to and the need to ensure that no onetrainee’s responses can be identified in the report;

● that the reporting tool allows deaneries to disseminate the findings to theirproviders with their local contextual information included;

● which aggregations are most important and should be reported on first;

● other desirable aggregations such as hospital based VTS GP trainees only;

● benchmarking groups – for each aggregation the units can be grouped indifferent ways for comparison purposes, for instance one could comparecardiology trainees’ experience with the experience of all other medicalspecialties, or just the trainees in the acute medical specialties.

When considering proposals for reporting it is important to remember that PMETBhas made the following guarantees:

● all survey data are confidential and it will not be possible for deanery staff toidentify the responses of an individual trainee doctor;

● data will be reported in such a way that responses remain anonymous and cannotbe identified due to small numbers in any particular group.

Population data

The 2006 survey’s population comprised all trainees in approved training posts.PMETB was dependent on deaneries to identify these doctors, as there is presentlyno single, centrally maintained UK database of trainees in approved posts.Headcount of doctors in approved posts by specialty, provider and grade was notavailable at the start of the data collection period due to time pressure, whichmeant that response rates could not be calculated immediately. In addition, therewere issues around how posts are defined as approved, since there are posts thatare deanery funded and educationally approved, posts that are locally funded andeducationally approved, and posts that are service only. Information on these postswas not generally held in one place; for instance, deaneries hold SpR data sincethey issue National Training Numbers, but deaneries only hold whole timeequivalent data for SHO posts that are deanery funded. Data quality issues wereapparent; for instance, over 200 of the same SpRs were listed on two differentdeaneries’ databases.

Page 54: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

54 National Trainee Survey 2006 – key findings

Both PMETB and the deaneries are committed to improving these data for the 2007survey. PMETB has been consulting with deaneries’ business managers over howbest to obtain data on the population of trainees in approved posts, and will beissuing their data request well in advance of the 2007 survey. At a national and UKlevel, PMETB is working with representative groups to discuss the development of aminimum data set, of which these data will be part. These proposals will beincluded in the consultation.

Deaneries will know how many specialty trainee posts there are, as these werecounted as part of the preparation for the implementation of Modernising MedicalCareers (MMC)23. However, the national counting of fixed-term specialty trainingappointment (FTSTA) posts may be an issue as, according to the forthcoming Aguide to specialty training: the gold guidexxvi, there is no mandatory requirement fordeaneries to record these posts and currently no specified data template on whichto do so. PMETB will consult with deaneries to establish an agreed template forrecording details of these posts as part of the QA framework.

In England, the Department of Health conducts an annual census of medical staffxxvii.It uses pay scales to classify doctors by grade and does not collect data on whetherthe doctor is in an educationally approved post. Trainees are currently classified as:

● registrar group, which refers to the combined grouping of specialist registrars,senior registrars and registrars and other staff working at equivalent grades thatare not in an educationally approved post;

● doctors in training and equivalents (previously known as junior doctors): registrargroup, senior house officer, foundation doctors, house officers and other staffworking at equivalent grades that are not in an educationally approved post.

PMETB has requested that the Information Centre for Health and Social Care inEngland include a field on the census to indicate whether the doctor is in aneducationally approved post or not. PMETB will also establish whether the otherthree nations’ departments conduct similar censuses of medical staff.

The survey as screening tool

The utility of the survey as a screening tool is determined by its predictive validity -whether it can identify posts or programmes that may require remedial action orhighlight areas of good practice. The forthcoming consultation on the QAframework will identify a criterion score that could be used to measure the survey’spredictive validity. A measuring instrument may have to be developed; scores fromthis can be used as the dependent variable, with the indicator scores from thesurvey as the independent variables. This will establish whether trainees’perceptions of a provider correlate with the criterion score. To avoid rangerestriction, the criterion score will have to be collected from providers at all levelsof the trainees’ ratings, some from the bottom quartile and some in the middle, andso on for the duration of the exercise. For the purposes of the validation exercise,anyone involved in obtaining the criterion score will need to be blinded to thesurvey data.

The exercise could establish which indicator scores are most effective and adjustthe identification of outliers to ensure that the survey is more sensitive than specific,as it may be better to have false positives than for PMETB or the deaneries to fail tofollow up a problem.

23 See: http://www.mmc.nhs.uk/pages/home

Page 55: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

National Trainee Survey 2006 – key findings 55

Making the survey mandatory

Feedback from deanery and postgraduate centre staff indicates that making thesurvey mandatory could make administration far easier. Due to the different routesof administration employed, it would not have been possible to enforce this for the2006 survey. The forthcoming A guide to specialty training: the gold guide and theGeneral Medical Council’s Good medical practicexxviii both indicate that doctors arerequired to take part in systems of quality assurance and quality improvement. Thesurvey forms a key part of the quality assurance of doctors’ training programmesand they will therefore be expected to complete a return for future surveys. As thesurvey will be exclusively administered through a website, which will providereference numbers for successfully completed submissions, doctors will be able toshow that they have completed the survey. In addition, PMETB will be able to listwho has not completed a return and supply this information to the deaneries andcentres for follow-up action. Together, this will provide a mechanism for deaneriesto encourage a high response rate.

Recognising that PMETB is committed to the principles of better regulation and sohas a duty to minimise the burden of its data collection activities, PMETB will ensurethat doctors’ participation in the work is as easy as possible and will thereforeprovide support throughout the survey period to ensure that doctors are able tocomplete a return. In addition, a full pilot of the process will be conducted in onedeanery to spot any process issues prior to national roll-out.

Comparison with 2006 data

Once data from two years are available it will be possible to perform analysislooking at the relationship between the 2006 and the 2007 indicator scores forparticular aggregations; for instance, for a given specialty group, do the samelocations come up as low scoring? In addition, the 2006 data provide a baselineprior to the implementation of MMC.

National use of the data by Royal Colleges

For the 2006 survey, PMETB supplied Royal Colleges with national levelaggregations for each specialty. These data will be made available using the 2007data too, but for 2007 these data will be of more interest as they will include thespecialty specific items that the colleges will have helped to develop.

The 2007 survey and other data sources

The survey data can be used to test other hypotheses. For example, if it is knownthat some trusts are participating in a quality improvement activity and otherswithin the deanery are not, a comparison of the relevant indicator score can bemade between the two groups. Analyses such as these can be performed, providingdeaneries submit the grouping data available and it is possible to specify whichindicator scores should vary by group to avoid multiple testing.

During the 2007 trainee survey, a survey of trainers will be conducted concurrently.The aggregated data from these two surveys will be linked using specialty andlocation. This will allow comparison of the trainees’ and trainers’ perspectives andthe potential development of discrepancy indicators; the greater the differencebetween the two perspectives, the greater the potential problem. It is likely thattrainees’ perceptions will not be as positive as those of trainers. Baker andSpracklingxxix compared the two perspectives (consultants and their VTS SHOs) and

Page 56: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

56 National Trainee Survey 2006 – key findings

found that consultants reported more positive perceptions. For instance, 32 per centof the trainees in their sample indicated that teaching took place in protected time,whereas the figure for consultants was 67 per cent.

Deanery action plans

To ensure that survey data are always used for quality improvement work, actionplans should be written for an external audience so that trainees can see that theirparticipation in the survey is a worthwhile activity, and that patients can bereassured that deaneries are effective in ensuring the safe training of doctors.PMETB will also be seeking evidence of trainee involvement in the agreement andmeasurement of the action plans.

Indicative plan for the 2007 survey

Table 24 shows key external dates for the trainee in the 2007 survey plan. PMETBrecognises that deaneries need sufficient notice that data will be released to allowthem to schedule resources for onward dissemination of the data to their providersand the preparation of action plans in conjunction with the providers. PMETB willprovide more detailed plans nearer the time.

Feedback form

If you have comments on the 2006 work or the plans for the 2007 work, please go towww.pmetb.org.uk/traineesurvey and complete the feedback form.

Milestone

Development of specialty specific items in conjunctionwith lead deans and Royal Colleges/faculties.

Population request to deaneries for completion. This willconsist of list of approved posts detailing the specialty,grade and location of each.

Population request returned to deaneries for them tosupply the email addresses of the posts’ currentincumbents. A draft version of this request is available atwww.pmetb.org.uk/traineesurvey.

Full pilot of trainee questionnaire in one deanery.

Staged launch of National Trainee Survey.

National Trainee Survey reports (all aggregations, seeTable 22) made available to deaneries, Royal Colleges,faculties and specialist associations.

Date

April – July 2007

June 2007

August 2007

October 2007

December 2007– January 2008

March 2008

Table 24 2007 survey plan

Page 57: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

National Trainee Survey 2006 – key findings 57

References

i PMETB. Generic standards for training (2006). Available at:http://www.pmetb.org.uk/fileadmin/user/Policy/Policy_Statements/Generic_standards_for_training_April_06.pdf

ii Grant, J, Maxted M, Owen H,Wooding K, Black C, Neville E, Gale R.Development Of A National Generic Data Collection System To SupportInspection Of Hospital Visiting (2004). Available at:http://www.pmetb.org.uk/index.php?id=backgroundinfo#c977

iii Paice E, Aitken M, Cowan G and Heard S. (2000). Trainee satisfaction beforeand after the Calman reforms of specialist training: questionnaire survey.British Medical Journal;320: 832-836.

iv Leach C and Goodman J. (2006). PMETB Questionnaire Assessment Final Report.Available at: http://www.pmetb.org.uk/index.php?id=backgroundinfo#c977

v Baldwin D C Jr and Daugherty S R. Sleep deprivation and fatigue in residencytraining: results of a national survey of first- and second-year residents. Sleep.2004 Mar 15;27(2):217-23.

vi Healthcare Commission (2006). Working in partnership: Getting the best frominspection, audit, review and regulation of health and social care.Available here: http://www.concordat.org.uk/_db/_documents/Concordat_-_version_(May_06).pdf

vii Bland M J, Altman D. (1995). Multiple significance tests: the Bonferroni Method.British Medical Journal. 1995;310;170.

viii Nunally J C. (1978). Psychometric theory (2nd edn). New York: McGraw-Hill.

ix Cronbach L J and Meehl P E. (1955). Construct Validity In Psychological Tests.Psychological Bulletin, 52, 281-302.

x Cohen J and Cohen P. (1983). Applied Multiple Regression/Correlation Analysisfor the Behavioral Sciences. (2nd ed.). Lawrence Erlbaum Assoc.

xi Goldstein H. (1995). Multilevel statistical models. 2 ed. London: Arnold.

xii McManus C. (2007). Multilevel modelling of PMETB data on trainee satisfactionand supervision. Report produced for PMETB available at:http://www.pmetb.org.uk/index.php?id=2006results#c978

xiii Hicks B. (2000). Time to stop bullying and intimidation. HospitalMedicine;61:428-31.

xiv Paice E, Aitken M, Houghton A, Firth-Cozens J. (2004). Bullying among doctorsin training: cross sectional questionnaire survey. British MedicalJournal;329(7467):658-9.

xv Altman D, Machin D, Bryant T N and Gardner M J. (2000). Statistics WithConfidence: Confidence Intervals and Statistical Guidelines. London: BMJ Books

xvi Frank E, Carrera J S, Stratton T, Bickel J and Nora L M. Experiences ofbelittlement and harassment and their correlates among medical students inthe United States: longitudinal survey. British Medical Journal;333;682-8.

xvii For details see: http://www.riti.org.uk/index5.html

Page 58: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

58 National Trainee Survey 2006 – key findings

xviii For details see:http://www.healthcareworkforce.nhs.uk/index.php?option=com_d4j_ezine&Itemid=332

xix Mahon A, Harris C, Tyrer J, Carr S, Lowson K, Carr L, Chaplin S and Wright D.(2005). The implementation and impact of Hospital at Night pilot projects: Anevaluation report. Department of Health. Available at:http://www.wise.nhs.uk/sites/workforce/usingstaffskillseffectively/Hospital%20at%20Night%20Document%20Library/1/Hosp%20At%20Night%20Final%20Report.pdf

xx Edwards P, Roberts I, Clarke M, DiGuiseppi C, Pratap S,Wentz R, Kwan I.(2002). Increasing response rates to postal questionnaires: systematic review.British Medical Journal;324;1183-1191.

xxi Training and Development Agency for Schools (2006). Results of the newlyqualified teacher survey 2006. Available here:http://www.tda.gov.uk/upload/resources/pdf/n/nqt%20survey%202006%20-%20sector%20level%20report.pdf

xxii Spector P E. (1994). Using self-report questionnaires in OB research: acomment on the use of a controversial method. Journal of OrganizationalBehaviour, 15, 385-392.

xxiii Judge T A. (1993). Does Affective Disposition Moderate the RelationshipBetween Job Satisfaction and Voluntary Turnover. Journal of AppliedPsychology, 78, 3, 395-401.

xxiv Mackinnon A, Jorm A F, Christensen H, Korten A E, Jacomb P A and Rodgers N.(1999). A short form of the Positive and Negative Affect Schedule: evaluation offactorial validity and invariance across demographic variables in acommunity sample. Personality and Individual Differences, 27, 405-416.

xxv Merrill J M, Laux L F, Lorimor R J, Thornby J I,Vallbona C. (1995). Measuringsocial desirability among senior medical students. Psychological Reports. 1995Dec;77(3 Pt 1):859-64.

xxvi Department of Health (In press 2007). A guide to specialty training: the goldguide.

xxvii The Information Centre (April 2006). Hospital and Community Health ServicesMedical and Dental staff in England: 1995-2005. Available here:http://www.ic.nhs.uk/pubs/nhsstaff/medicaldentalintro/file

xxviii General Medical Council (2006). Good Medical Practice. Available here:http://www.gmc-uk.org/guidance/good_medical_practice/GMC_GMP.pdf

xxix Baker M and Sprackling P D. (1994). The educational component of seniorhouse officer posts: differences in the perceptions of consultants and juniordoctors. Postgraduate Medical Journal. 1994 Mar;70(821):198-202.

Page 59: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300
Page 60: National Trainee Survey 2006 – key findings · used by Royal Colleges and deaneries; its core questions were developed and validated through face-to-face interviews with over 300

Postgraduate Medical Educationand Training Board

Hercules HouseHercules RoadLondon SE1 7DU

Tel +44 (0)20 7160 6100

Fax +44 (0)20 7160 6102

www.pmetb.org.uk

© PMETB 2007

ISBN: 978-0-9555910-0-6