- Joined
- Aug 11, 2013
- Messages
- 779
- Reaction score
- 387
If the MCAT is a CBT, why the hell does it take a month to grade???? You would think the results could be generated in seconds before you leave the testing center....
Wow, I...I agree with you.If the MCAT is a CBT, why the hell does it take a month to grade???? You would think the results could be generated in seconds before you leave the testing center....
That seems counter to everything I've heard, but then, it is all *TOP SECRET* so what the hell do any of us know?I've heard they adjust the curve, throw out certain questions, and do some other shady ****. @$$holes
AAMC, get your act together or we'll take our business elsewhere! No we won't. We're sorry. We'll do whatever you say and take whatever you give us. Please don't be mad. 🤔
Why can't some other company compete and make a better MCAT that u get results immediately. Kinda like what happened when the ACT started to become more popular than the SATs.
Doesn't that kind of defeat the purpose of a standardized exam?
I don't think so. Colleges accept both SATs and ACTs. Doesn't seem to be an issue for them.
If the MCAT is a CBT, why the hell does it take a month to grade???? You would think the results could be generated in seconds before you leave the testing center....
I was wondering the exact same thing! My friends who took the GRE knew their results right away.If the MCAT is a CBT, why the hell does it take a month to grade???? You would think the results could be generated in seconds before you leave the testing center....
The scale is established based on previous tester performance on the scored questions, and experimental items are obviously known beforehand and will not be scored. It has nothing to do with "the way the curve is generated" and there is no reason for them to throw out questions based on any "curve" because there isn't one. Jepstein's post is correct, the scale (raw to scaled scores) is set beforehand based on the item composition of the test.It has to do with the way the curve is generated. Also, yes, they do throw out some questions based on the curve and because some are experimental. You Gen-Y-ers and your need for instant gratification 🙄...
So what is the reason for this thread besides bitching and complaining..?
I expect that they look at the scores and distributions on the whole test, the different sections, and individual questions. If everyone gets a question wrong, they might take a look at it and throw it out, or weight it differently, or something. I would expect and hope that they do a full review of the test given the scores and percentages. I was involved in some standardized test grading very briefly back in the day and there are a lot of little details and nudges put into the final score calculations.
It used to be because of the essay. I didn't expect them to completely do away with the waiting time, but I thought the 2015 MCAT may feature instant scoring. Not really sure why they don't as other graduate level tests do.
This is what I was referring to, though I realize the way it was stated in my previous post was incorrect. When I first started preparing to study for the MCAT (2009-2010ish) I had some info from the AAMC that corroborated with this. I've since gotten rid of most of my MCAT stuff, but I'll take a look and if I still have it I'll post it. From the AAMC's "Understanding Your Score Page:
"Why are raw scores converted to scaled scores?
The conversion of raw scores to scaled scores compensates for small variations in difficulty between sets of questions. The exact conversion of raw to scaled scores is not constant; because different sets of questions are used on different exams." This suggests that there is some post-test statistical analysis before scores are known.
Is the exam graded on a curve?
Examinees often ask if earning a high score or higher percentile is easier or harder at different times of the testing year. They ask whether they have a better chance of earning a higher score in April or in August, for example. The question is based on an assumption that the exam is scored on a curve, and that a final score is dependent on how an individual performed in comparison to other examinees from the same test day or same time of year.
While there may be small differences in the MCAT exam you took compared to another examinee, the scoring process accounts for these differences so that an 8 earned on physical sciences on one exam means the same thing as an 8 earned on any other exam. The percentile provided on your score report simply indicates what percentage of examinees from the previous testing year scored the same as you did on the MCAT exam.
How you score on the MCAT exam, therefore, is not reflective of the particular exam you took—including the time of day, the test date, or the time of year—since any difference in difficulty level is accounted for when calculating your scale scores (see above for information about scaling).
No, that means based on the set of questions, a scale is generated.
It's very simple.
1) AAMC* designs tons of MCAT questions of varying difficulties
2) AAMC puts these in MCATs as experimental (unscored) questions to get data on how current test takers perform on these questions
3) Based on that data, AAMC designs an MCAT that has roughly the same difficulty as all the other MCATs ever offered
4) Because it's troublesome to get the exact same difficulty level, the AAMC can also slightly adjust the conversions to do the job for them.
The MCAT is not curved. The MCAT does not throw out any questions if everyone got them wrong. All of the questions you get were already taken by hundreds of test takers.. the ones that are 'unfair' have already been weeded out.
Seriously, look up the difference between 'scaled' and 'curved'. Two very different things.
From that same page,
*and by AAMC, I mean the testing company they hire to produce questions
I don't understand why post-test would be more accurate/valid. If anything, it is far MORE valid to base them on a large base of students taking the MCAT over multiple different sittings, rather than the group that happens to go in August. Hey, maybe the students taking it in August are more prepared because they spent all summer studying. If you were to give more credence to a post-test analysis for that group, you would essentially take your scaling and turn it INTO a curve, which would make the grades a lot less consistent between sittings and a lot more "I am being compared to all of the students who studied over the summer if I take in August, I should take it in March so I'm competing against all of the last-minute peeps who can't afford to delay any longer even if they're not prepped".I understand the difference between a scaled test and curved test, and I've already acknowledged that I misspoke about this in a previous post. Regarding your fourth point, for these adjustments to be as accurate and valid as possible, it would make the most sense to do them after a post-test analysis. Doing so pre-test would require assumptions and leave room for error.
I don't understand why post-test would be more accurate/valid. If anything, it is far MORE valid to base them on a large base of students taking the MCAT over multiple different sittings, rather than the group that happens to go in August. Hey, maybe the students taking it in August are more prepared because they spent all summer studying. If you were to give more credence to a post-test analysis for that group, you would essentially take your scaling and turn it INTO a curve, which would make the grades a lot less consistent between sittings and a lot more "I am being compared to all of the students who studied over the summer if I take in August, I should take it in March so I'm competing against all of the last-minute peeps who can't afford to delay any longer even if they're not prepped".
You may understand the difference between a scaled grade and a curve, but you seem determined to turn the scaling system INTO a curve.
2+2 = 4 whether you did a multiplication problem before it or a division. Each passage is independent; my performance on one isn't going to be affected by the other passages in an exam.Because of the context in which the questions are presented, which is a significant factor in developing any type of metric. It's the first time all of those questions will be presented and evaluated together in that context. One possible effect of this is, as previously mentioned by jepstein, is inadvertently creating a more difficult and taxing exam. It may seem arbitrary but it can have a significant effect on a metric's validity. Thus, pre-test they may have a good idea of where adjustments should be in conversion factors, but the only way to confirm this is with post-test analyses.
Just some theories:
I have not seen any indications that the correct answers are even available in the test center. It is quite possible that AAMC distributes only the questions, collects the answers and grades them later. Considering that leaving the test computers connected to the internet is not the wisest move and that the CBT has been developed some time ago, I would not be surprised if the data from the test centers is sent to AAMC on some sort of physical carrier.
Has anybody else notice when you take a AAMC practice exam you get your score in seconds....
Yes. Give them a preliminary, unofficial score right away with full knowledge that it's not real until the score report comes through.I'm sure it's not totally unheard of for a question to be thrown out or re-graded for difficulty after the fact because of something or other, even if just a typo. I'm sure it's also not totally unheard of for a test center irregularity to be discovered, or some error to have been made.
Imagine if an immediate score was given, and it turned out a passage had a typo that resulted in many of the answers being wrong.. but everyone had their scores when it was discovered. Or if there was a typo in the answer key that resulted in a bunch of people who chose the wrong answer but got it marked correct. What do you do? Show them "Your score: 34" at the end of the test then send them a letter 3 weeks later saying "just kidding, you actually got a 33"?
No, that's a curve. See the discussion above; AAMC scales the scores, they don't curve. It's slightly different.Ever notice how the grading rubric for different practice tests are sometimes different? (E.g. 22-23 correct answers = 10 on one test while it may be 23-24 = 10 on another).
AAMC waits to grade so that they can evaluate and compare how EVERYONE did on their test before modifying the score. They score by percentile, not by simply how many questions you got correct. In order to do this, they need everyone to finish taking the test, else it's just an educated guess.
At least this is what I have always thought. I could be wrong, then again I don't care enough to delve any deeper.
I'm sure it's not totally unheard of for a question to be thrown out or re-graded for difficulty after the fact because of something or other, even if just a typo. I'm sure it's also not totally unheard of for a test center irregularity to be discovered, or some error to have been made.
Imagine if an immediate score was given, and it turned out a passage had a typo that resulted in many of the answers being wrong.. but everyone had their scores when it was discovered. Or if there was a typo in the answer key that resulted in a bunch of people who chose the wrong answer but got it marked correct. What do you do? Show them "Your score: 34" at the end of the test then send them a letter 3 weeks later saying "just kidding, you actually got a 33"?
your an idiot
I understand the difference between a scaled test and curved test, and I've already acknowledged that I misspoke about this in a previous post. Regarding your fourth point, for these adjustments to be as accurate and valid as possible, it would make the most sense to do them after a post-test analysis. Doing so pre-test would require assumptions and leave room for error.
Ever notice how the grading rubric for different practice tests are sometimes different? (E.g. 22-23 correct answers = 10 on one test while it may be 23-24 = 10 on another).
AAMC waits to grade so that they can evaluate and compare how EVERYONE did on their test before modifying the score. They score by percentile, not by simply how many questions you got correct. In order to do this, they need everyone to finish taking the test, else it's just an educated guess.
At least this is what I have always thought. I could be wrong, then again I don't care enough to delve any deeper.
2+2 = 4 whether you did a multiplication problem before it or a division.
Each passage is independent; my performance on one isn't going to be affected by the other passages in an exam.
If it's really that much of an issue, give them a range: there is a 95% chance (base that number on the stats of how often these things actually change from prelim to final) that you scored between 32-34.
That's great and all.. but it's not how they do it.
And you know this how? The AAMC keeps the methods used for scaling raw scores confidential, its anyone's conjecture. Logically, for accuracy in determining scaled scores and for the delay in reporting scores, post-test analyses makes much more sense.
Every form of the MCAT exam measures the same basic skills and concepts. However, each form is different in regard to the specific questions it uses. Because each form has the potential to be easier or slightly more difficult than another, raw scores are converted to a scale that takes into consideration the level of difficulty of the test questions on a given form. This conversion minimizes variability in the meaning of test scores across forms
You can absolutely make confidence interval based on, statistically, how often scores change between prelim and final. They can GET actual data for that very, very quickly, even if they aren't willing to go into the HUGE pile of data they already have sitting around.You can't really make a truly accurate confidence interval without the actual data. I'm not denying that it sucks to have to wait for a month for your results, but there is a perfectly valid reason for it. I know its not the most fair comparison, as the turn around time is much shorter (though I can think of several exceptions from my own experience), but professors don't give you an idea of how you scored on a test as soon as you hand it in. Aside from grading it they have to see if there were any issues, and in some cases run some stats on the results before they're final (I've had several profs. who did these). Why some other test should be held to a different standard to appease some people's neurosis is kind of ridiculous IMHO.
Based on information readily available on AAMC's own website and general knowledge of how the MCAT works? Plus info learned from teaching for the MCAT through a large test prep company?
Also, by the very fact that its a SCALE not a curve. You still don't seem to know the difference.
So we have the exact same knowledge base, but I choose to look at the unknown aspect of how the scaled scores are derived from a perspective that would yield greater accuracy.
I unintentionally misused curve once and admitted it in my next post. I haven't misused scale or curve since. So your point is...