Went away for the weekend. Lots to get caught up on.
If P/F for S1 were to come to pass, programs would need to make decisions using other factors. S2 might be used, forcing students to take it earlier. I don't think we're ever going back to a world where there is no standardized exam to take. The big fields will make their own, if that's what's needed.
There is no clear timeline. The "plan" is to convene some sort of a group to review all of these issues, and (theoretically) make a decision in 2019-2020. But getting consensus is going to be difficult, if not impossible. Even if a decision is made in this timeframe, any change would be delayed for a few years. This isn't worth worrying about.
Answered by some others already, but depends on what you mean by "holistically". MSPE's do stratify students into groups. LOR's are not very helpful -- all say the student is in the top 5%, and relying on LOR's becomes a "who you know" / "How connected you are" game. So all we're left with are clinical grades and MSPE.
And, that's for US grads. 50% of IM spots are filled by IMG's, and many of them have useless MSPE's. All they have are their USMLE scores.
One of the options was to divide into quartiles or quintiles. It would be fine as far as I'm concerned. Problem is, if you're in that bottom group, I can't tell if you just barely passed, or if your performance was better than that. Some programs will assume the worst.
We've had this discussion before, but 1) that might not be possible, and 2) this just elevates the importance of other parts of the application which may be just as flawed, or worse. Plus, IMG's fill a large number of spots in the US, and their applications may be mostly worthless except for exam scores. And if we just switch to S2 scores, we're just pushing the problem onto a new exam (and students will want "dedicated" 8 weeks to study for it).
This study often comes up in this discussion. It was done looking at graduates in 1997-2000, when step scores were lower (av S1 was 215). The USMLE "states" that scores are equivalent over time. Either that's wrong and there's some inflation over time, or we licensed a bunch of people in the past who should have failed (as the minimum pass on S1 has risen from 175 to 194 -- you can see the trends here:
A peek inside the USMLE sausage factory: setting the Step 1 minimum passing score). I think it's the former.
So, problems with this study:
1. If 215 => 93% board pass, that's a 215 in 1997 which was the average. It certainly isn't a 215 today.
2. A 93% board pass rate would put an IM program in the 3rd or 4th quartile (bottom) of IM programs. That is a terrible board pass rate.
So it's an interesting study, but limited.
This is one of the options listed. Would students be required to take both? Or could they just take S1 and if "happy", just plan to take S2 later (much like today)? And since S2 scores are higher than S1 scores on average, this will just lead to further score inflation.
It's a tradeoff, and they are trying to have it both ways. Since CS is P/F, if you fail you can't show how badly you failed (just barely vs miserably), and then when you subsequently pass can't show whether you passed barely or by a mile. Adding scoring would help that, but then we're back to S1 mania issues with scores.
It's an enormous problem. Everyone is in the top 5% in LOR's. Many MSPE's are at least more reasonable, but some continue to place students into "groups" but then don't tell you how many are in each group.
Either is a reasonable option. I agree the first will drive students crazy trying to determine what number of applications is "best". And the latter, some fields may decide to only review the preference apps, placing a defacto hard limit on applications and forcing students to decide which programs they are really competitive for.
The explorer tool may end up being useful, probably better than where we are now.
Sure, but then your entire career hangs in the balance of a single evaluation. In the ED you work with a mix of people so perhaps it's a bit more fair. And schools would need to ensure that not 95% of student get the top rating. Soon students will be complaining that the evaluations are subjective and unfair, and that they want an objective system. Plus this create a situation where you only get your eval when your training is just about done -- if it's poor, you have little time to reboot your career.
And, not useful for IMG's. So really not a system wide answer.
Might be taking your comment out of context (sorry!), but this is also a common statement -- that because the SD is wide enough that most scores can't be compared. This is not exactly accurate. First, these statistics are used to compare the averages of two groups to see if they come from the same population, not comparing two individual scores. Second, this all assumes a 95% confidence interval -- I'm happy with a much lower level of confidence. The number quoted (16?) - not sure where that is from. If it's the standard dev of the S1 score distribution, that's not what you want.
My bottom line:
1. If the FSMB and the AAMC want to change to P/F, there's no one to stop them.
2. As they seem to have punted this to some still-to-be-determined group of concerned citizens, I doubt they will be able to come to any consensus.
3. As a non trusting sort, I would move ahead with development of a specialty specific exam so that, as a field, we simply don't need to worry about this any more. Likely outcome is --> group formed --> lots of talk --> no agreement --> AAMC/NBME/FSMB decide whatever they want anyway.