<p>The last time you posted this exact same thread, davida1, I said the exact same thing.</p>
<p>Your ranking is absolutely worthless. </p>
<p>To begin with, it makes absolutely no sense to rank by absolute numbers. If Penn is twice as large as Dartmouth but sends the same number of students to grad schools, that doesn't show that the schools are doing equally well; it shows that relatively Dartmouth is doing twice as well. If you're worried about how Cornell's school of Hotel Management affects the numbers, figure out the relative %age of Cornell students attending the top grad schools assuming none are coming out of that school. The numbers will now favor schools like Cornell (because obviously some people out of the Hotel Management school do go on to top grad schools), but that is clearly preferable to numbers that reflect size more than quality of an undergraduate institution.</p>
<p>Now, we are left with serious problems with your data collection mechanism. Facebook itself is self-selecting, and facebook groups are even more so. It's clear that only a fraction of those actually attending the schools you're measuring join the facebook group for the school; Yale, for example, has 30% fewer students signed up for facebook groups at the 45 schools you're measuring (which include all of the WSJ's schools) than at the 15 schools measured by the WSJ. However, Cornell has a 10% increase. Does this indicate that relatively more students at Cornell go to the top grad schools, or does it indicate that relatively more students at Cornell join facebook groups for professional schools? Alternatively, it is possible that Yale students are more likely to join facebook groups than Cornell students, and in fact, Cornell's increase should be 100% and Yale's decrease should be 60%. The fact is, we don't know and can't make an educated guess about anything based from your numbers. </p>
<p>In slightly more scientific terms, your numbers fail to be meaningful because of selection bias; you are not measuring a random group of students accepted at top schools, but rather, a group of students who have elected to join facebook groups for top schools. </p>
<p>Now, as I've said earlier, your problem is fairly easily adjusted for. We have the data points we need to approximate what the facebook self-selection bias may be in a given school, and to adjust for it. If we find the number of students in facebook groups for the exact same selection of schools as WSJ uses and compare those numbers, undergraduate school by undergraduate school, we can approximate the selection bias for facebook groups on a school-by-school basis. Then, it will be easy to apply that bias to the per-capita numbers that you have found. That should yield a result far more indicative of reality. </p>
<p>For example, if there are 100 students from Berkeley in facebook groups for the WSJ's selection of feeder schools, but WSJ lists 200 students attending these schools, then we can assume that the Berkeley facebook selection factor is approximately 50%. Thus, if there are 280 Berkeley students enrolled in your selection of feeder schools, we can assume that this represents roughly half of the total Berkeley students at these grad schools leaving us with an adjusted number of 560 Berkeley students at grad schools. </p>
<p>Davida1, if--as you say--it was easy to collect this information, it should not be difficult for you to make these adjustments. If--as you say--you are motivated by a desire to rank feeder schools accurately (rather than, say, a bias towards one school in particular), I am sure that you will make these adjustments. Without performing these calculations, your data is useless even for making the most basic of assumptions.</p>