Two radiologists take vastly different paths reviewing a case - one far more thorough than the other. But what’s the payoff for the extra work? Is it enough?
For those seeing the title of this piece and becoming concerned: Rest assured that I am not about to suggest sealing a rad in a box (no matter how similar your reading room might be to one) or subjecting him to a random chance of extermination.
Rather, consider a hypothetical imaging study, just performed and about to be assigned to a radiologist’s work list. Radiologist A or his colleague, B, of equal skill and experience, will get the study for interpretation, based on the vagaries of their shared work lists and PACS. I do not pretend to understand quantum mechanics well enough to further stretch the analogy. If it helps you, pretend this is a Twilight Zone episode.
Radiologist A gets the case: Sizes up the provided clinical history, looks over the study, activates his macro for a “no significant pathology” case, and moves on. Took him a couple of minutes.
Radiologist B gets the case: Sees the same clinical history but feels it is inadequate for diagnostic and/or reimbursement purposes. Manages to track down someone on the clinical service who can supply better background. In the process, notices that the patient had prior studies which were not uploaded to the system, and after getting off the first phone call rings up the technologist to reload the prior. Tries looking over the images while distracted by these calls, noticing nothing emergent but sets the case aside until the comparison can occur because there are a couple of uncertain issues (possible 1 mm pulmonary nodules, ditzels in the liver, etc.).
B gets the call from the tech that the prior is now loaded, but it’s over half an hour later and his recollection of the case is already fuzzy, so he essentially rereads the study as he compares with the previous case. Recognizes that the 1 mm incidentalomas are stable, but now sees that the previous report did not mention them. Agonizes a while about how to phrase his dictation so as not to unduly worry the patient or referring clinicians, also to avoid putting the previous study’s reader in a bad medico-legal spot.
Finally ready to sign the report, Rad B notices that the referring clinician has been flagged as a doctor who wants to be personally called on all cases, positive or negative. Sighs, picks up the phone, and engages in a fruitless series of attempts to contact the referrer by every means he can conceive, without success. Agonizes further about how to document these efforts so as to avoid the appearance of noncompliance with department regulations, yet also avoid documenting the clinician’s lack of responsiveness in the legally-discoverable written record.
Radiologist B has spent nearly an hour with various aspects of this case, not counting the additional time he will endure a nagging feeling that some aspect of the thing will come back to haunt him during the next few days. Remember that his colleague, A, spent a couple of minutes, and if you asked him about the case a week, a day, or an hour later, he’d have to look it up to know what you were talking about.
Like the cat of the original analogy, a patient whose case might go to radiologist A or B, knowing how they’d handle his diagnostic care, would likely have a strong preference in the matter. Physicians referring patients to A and B’s radiological group, the group itself, and the society collectively paying for and depending on healthcare surely would as well.
Maybe Radiologist B has a 0.1 percent lower chance of getting named in a lawsuit as a result of his more thorough, conscientious approach. Maybe his peer-review stats will look a bit nicer. Maybe he hopes to establish a good track record that will give him a slight edge over A for job security, promotion, etc. in the coming decade - assuming A’s greater productivity numbers don’t get the spotlight instead.
Is this the sum total of what we’re doing to encourage B to keep it up, and not become more like A in the future?
Emerging AI Algorithm Shows Promise for Abbreviated Breast MRI in Multicenter Study
April 25th 2025An artificial intelligence algorithm for dynamic contrast-enhanced breast MRI offered a 93.9 percent AUC for breast cancer detection, and a 92.3 percent sensitivity in BI-RADS 3 cases, according to new research presented at the Society for Breast Imaging (SBI) conference.
The Reading Room Podcast: Current Perspectives on the Updated Appropriate Use Criteria for Brain PET
March 18th 2025In a new podcast, Satoshi Minoshima, M.D., Ph.D., and James Williams, Ph.D., share their insights on the recently updated appropriate use criteria for amyloid PET and tau PET in patients with mild cognitive impairment.
Can Abbreviated Breast MRI Have an Impact in Assessing Post-Neoadjuvant Chemotherapy Response?
April 24th 2025New research presented at the Society for Breast Imaging (SBI) conference suggests that abbreviated MRI is comparable to full MRI in assessing pathologic complete response to neoadjuvant chemotherapy for breast cancer.
Clarius Mobile Health Unveils Anterior Knee Feature for Handheld Ultrasound
April 23rd 2025The T-Mode Anterior Knee feature reportedly offers a combination of automated segmentation and real-time conversion of grayscale ultrasound images into color-coded visuals that bolster understanding for novice ultrasound users.