According to the JAMA Surgery paper, faculty evaluations encompassed technical skill and the six core competenciesmedical knowledge, patient care, interpersonal and communication skills, professionalism, practice-based learning and improvement, and system-based practice.
The paper analyzed data for 150 residents at different levels of training over 4 years and also found that even faculty evaluations of the category medical knowledge couldnt predict who would get a good or a bad score on the test.
Its great to know that at the authors institution, the average annual evaluation scores ranged from just over 75 to 100 with means and medians both slightly above 92like Garrison Keillors mythical Lake Wobegon, where all the women are strong, all the men are good looking, and all the children are above average.
Medical knowledge can be measured, but the other parameters are so subjective that they border on meaningless. They remind me of the infamous smiley face numerical pain scale that means different things to different patients.
Some examples. Earlier this year, I wrote about the difficulty defining professionalism. Using a numerical scale, how can you rate one resident as more professional than another?
And I always had trouble ranking one resident over another in system-based practice. It might be better to rate system-based practice on a binary scale; that is, can a resident define the term or not?
Big business is having trouble evaluating employees too. The evaluation process at General Electric was examined by Quartz. At GE, the annual review is not effective for managing people or improving performance. It leads to a tendency to focus excessively on process over outcomes and is an exercise in paperwork and bureaucracy instead of an agent of change.
Note that the JAMA Surgery study accumulated 1131 evals. Even if that was only virtual paperwork, its much work for little value, but at least there was a lot of data to show a site visitor from the Residency Review Committee.
A New Yorker article noted that consulting firm Deloittes evaluation process involves consensus meetings ending with managers marking on a 5-point scale how strongly they agree with two statements: Given what I know of this persons performance, and if it were my money, I would award this person the highest possible compensation increase and bonus; and Given what I know of this persons performance, I would always want him or her on my team. And they must answer yes or no to two more: This person is at risk for low performance, and This person is ready for promotion today.
Maybe we should adopt a modification of Deloittes system for our resident evaluations. Faculty must respond yes or no to this statement: I would let this resident operate on me. If the answer is no, why should we let that resident operate on anyone?
This post originally appeared on Physicians Weekly.