Wednesday, June 03, 2009

Ranking News

The European Union is planning on introducing a rival to the Shanghai and THES-QS rankings. This is a good idea in principle but who is going to get the contract? It is a pity that "internationalisation" is going to be an indicator and what exactly does "community outreach" mean?

Odile Quintin, the European Commission's director-general for education, told the HES that the Shanghai Jiao Tong was "firmly concentrated on research", anchored to the production of Nobel laureates, and narrow in scope.


"We think that universities have a strong role in research but also in teaching and employability so we are promoting an alternative ranking to measure all these dimensions," she said.


The ranking would be handled by a consortium working independently of the EC, and work would begin after the results of a tendering process were revealed next week.


The plan is to develop the ranking throughout 2009 and 2010, for implementation a
year later. The project will have a budget of E1.1 million ($1.9m).


Ms Quintin said the new ranking, while based in Europe, would have a global reach.
She added that the new European survey would be focused much more on
disciplinary strength, "because you can be the best university in nanotechnology
but not in psychology".

She said the alternative world ranking system would be independent, run neither by governments nor universities and provide a multidimensional measure of education, research, innovation, internationalisation and community outreach.

Tuesday, June 02, 2009

Saturday, May 30, 2009

Asian University Rankings

QS Quacquarelli Symonds has come out with a ranking of the top 200 Asian universities. Here is the top ten.

1. University of Hong Kong
2. Chinese University of Hong Kong
3. University of Tokyo
4. Hong Kong University of Science and Technology
5. Kyoto University
6. Osaka University
7. Korean Advanced Institute of Science and Technology
8. Seoul National University
9. Tokyo Institute of Technology
10. National University of Singapore and Peking University


There are also rankings by disciplinary cluster and by indicator.

For every single disciplinary cluster, the University of Tokyo, not the University of Hong Kong is top. How strange.

For the indicators, the National University of Singapore is first for Employer Review and International Students, Tokyo University for Academic Peer Review, College of Medicine at Pochon Cha University (Korea) for faculty student ratio [I’m wondering about that as well], Gwangju Institute of Science and Technology for papers per faculty, Yokohama City University for Citations per Paper, Hong Kong University of Science and Technology for International Faculty, Kansai Gaidai for Inbound Exchange Students and City University of Hong Kong for Outbound Exchange Students.


These rankings seem to be a shrewd marketing move. Universities that have no chance of getting anywhere in the World University Rankings will now be able to boast that they came in the top 50 Asian universities for outbound exchange students or top 100 for citations per paper. A glance at the indicator rankings, for example, shows some Malaysian universities that one would not have thought had any chance of being in any sort of ranking. On the other hand, these rankings have been able to identify rising stars such as the Multi Media University.

There are two methodological innovations, both of which are questionable. They need to be discussed since this regional ranking could be a tryout for the global rankings. The first is the addition of two further measures of internationalization, inbound and outbound exchange students.

If internationalization is going to be a criterion, then having more measures might be a good idea. However, it is time to consider whether internationalization is actually a valid indicator of quality. Measures of internationalization do not correlate very well if at all with any other indicator and they also give an unfair advantage to the European Union and Hong Kong.

If we want to measure faculty quality, which internationalization supposedly underlies, a better method might be calculate the percentage of a random sample of teaching and research staff on university web pages who obtained degrees from the top 100 universities (on the Shanghai rankings?).

However, since QS get a lot of their bread and butter from facilitating students moving across national boundaries we are unlikely to see the end of this indicator.

The addition of number of inbound and outbound exchange students might also be very easily manipulated. If it were included in the world rankings it is likely that we will see universities setting up branch campuses a few miles away across some increasingly irrelevant frontier and then moving everybody there for their second year and calling them exchange students. So we might expect to see Queens University Belfast setting up a branch in Dundalk in the Irish Republic or the National University of Singapore in Johore in Malaysia and so on.

The other innovation is that research is measured by citations per paper, which measures the average impact of papers, and papers per faculty which measures the quantity of research in a very basic sense. This represents an improvement over the previous policy of using a single indicator. However, the problem remains that both are based on the Scopus database which aims to be as inclusive as possible. Scopus is an excellent research tool but inclusion in its database is an indicator of quality only in the broadest sense. To be credible, QS should consider finding some measure of research that measures genuine excellence.

These rankings have some surprises, the most noticeable and one lacking in face validity, is that the University of Hong Kong and not the University of Tokyo is the top university in Asia. Or perhaps this should not really be a surprise. Tokyo actually outperforms Hong Kong on all indicators except the internationalization ones and is ahead in all of the disciplinary rankings. Again, a lot of South Korean universities do very well.

It is good that QS are prepared to experiment with different indicators but the methodological innovations of these rankings do not seem to help very much.

Wednesday, May 13, 2009

Jawaharlal Nehru University has been returned to India. (see previous post)

Tuesday, May 12, 2009

Don't They Teach Geography Any More?

The QS Asian University Rankings are now out. I hope to comment in a while. For the time being. I've noticed that in the "International Students Review" (I'm not sure that it's a review but never mind), Jawaharlal Nehru University is listed as being in South Korea? I'm wondering how long it will remain there?
Asian University Rankings

QS will shortly release their Top 100 Asian Universities Rankings. It seems from the bits and pieces released so far that Australia, the Pacific and South West Asia are not included. There appear to be two innovations -- a trial run for the global rankings? -- namely counting student exchanges and including citations per paper as a measure of quality of research.

Monday, May 11, 2009

The Man with the Midas Touch

Those who are familiar with Malaysian gymnastics – admittedly not a large group -- will know who I am talking about.

There is a gymnastics coach in Malaysia who has achieved remarkable results with the women’s team of a very small state. For the last few years this state has won gold after gold at national competition. Hence the title bestowed by the Malaysian press. There does not seem to be anything special about the state – the men’s team has never done very much. Nor is there anything unusual about his training methods apart from their rigour and his intolerance for poor performance.

There must then be something about his selection methods. The state is so small that it does not have a gymnastics association and so this coach is free to select anyone he wants for training.

Recently, there was an opportunity to see just what these selection methods might be. The coach has now moved up to the national team and held a selection for young gymnasts, one of whom was my daughter, to train at the national sports centre.

The selection was quite revealing. The gymnasts were required to run, jump, do a bridge, do as many chin-ups as they could and a few simple exercises. Their height and weight were measured. There was no interest in their competition records, team spirit, motivation, leadership qualities or ability to respond to adversity.

At the end there was an interesting moment. The gymnasts were told to line up with their parents behind. There was much scurrying into position as people assumed they were being summoned for a group photograph, something without which no Malaysian event of any sort is complete. But after surveying the lineup, the coach said thank you and waved everyone away. What he was in fact doing was checking to see what the gymnasts would look like in a few years

So that was the secret of the man with the Midas touch. Assessment of basic physical skills and characteristics and reference to inherited traits. It was as though someone selected for elite universities by a simple test of general intelligence and a check on parental academic performance.

So here are two proposed experiments. Asian universities – and others – should scrap the proliferating complex of personality tests, language tests, co-curricular activities, interviews to test for leadership, politeness, sensitivity, appearance, interest and so on, profiles, course work and essays and just test for general cognitive ability.

The second experiment is that this coach and others should take heed of the global consensus and introduce personality tests, interviews and the whole paraphernalia of holistic assessment to choose future athletes.

I have a horrible feeling that the first will never happen but that one day the second will.
Comments on an Article by Andrew Oswald

I have already referred to an article by Andrew Oswald (14/12/2007) that contained some acerbic comments on claims to excellence by British universities derived from their showing on the THE-QS rankings.

He recently had an article in the Independent suggesting three measures that would allow universities to rise in the rankings. I will skip the first and third and just look at the second. He proposes that universities should be free to pay the market rate for highly rated researchers and offers Dartmouth College, which built up a top economics faculty by paying outstanding researchers appropriate salaries.

Dr Oswald is quite right but a lot of people are going to be depressed after reading his article. How can universities in Africa, Latin America and Asia get anywhere if they cannot afford to pay Ivy League level salaries? Is there anything that a university without a big pot of money can do?

The answer is that there is. In Moneyball Michael Lewis described how the Oakland As baseball team performed dramatically well even though they had only a comparatively small amount of money with which to buy players. They did it by simply ignoring the intuitions of talent scouts and looking at crude statistical data.

Basically, Oakland did the equivalent of a university scrapping search committees, interviews, personality tests, references and looking at the research done by applicants. Or choosing undergraduates by testing cognitive skills, literacy and numeracy rather than holistic assessment of leadership, communication skills, response to adversity, community involvement and so on.

At least one American university has done something like this. George Mason University has built up an excellent economics department by recruiting academics specializing in unfashionable fields that were undervalued by the academic marketplace.

Asian universities and others might consider systematically recruiting researchers whose personal characteristics and choice of unpopular research topics put them at a disadvantage when applying for academic positions. They might end up with a collection of unpleasant eccentrics but they might also see their ranking scores inching upwards.

Sunday, March 29, 2009

The Globalization of Cheating

There is a fascinating story in the Chronicle of Higher Education about the growth of the essay writing business. It seems that around the world there are shadowy companies employing dozens or hundreds of writers churning out essays for university students at every level.

'The writers for essay mills are anonymous and often poorly paid. Some of them crank out 10 or more essays a week, hundreds over the course of a year. They earn anywhere from a few dollars to $40 per page, depending on the company and the subject. Some of the freelancers have graduate degrees and can write smooth, A-level prose. Others have no college degree and limited English skills.


James Robbins is one of the good ones. Mr. Robbins, now 30, started
working for essay mills to help pay his way through Lamar University, in Beaumont, Tex. He continued after graduation and, for a time, ran his own company under the name Mr. Essay. What he's discovered, after writing hundreds of academic papers, is that he has a knack for the form: He's fast, and his papers consistently earn high marks. "I can knock out 10 pages in an hour," he says. "Ten pages is nothing."
His most recent gig was for Essay Writers. His clients have included students from top colleges like the University of Pennsylvania, and he's written short freshman-comp papers along with longer, more sophisticated fare. Like all freelancers for Essay Writers, Mr. Robbins
logs in to a password-protected Web site that gives him access to the company's orders. If he finds an assignment that's to his liking, he clicks the "Take Order" button. "I took one on Christological topics in the second and third centuries," he remembers. "I didn't even know what that meant. I had to look it up on Wikipedia." '



There are interviews with some of the professional essay writers. Two of them are Americans with law degrees. Another is a Nigerian with a master's degree from the University of Lagos. Many others appear to be from the Philippines and India.

Customers of the essay mills include a doctoral student in aerospace engineering at MIT and graduate students at Northern Kentucky University, James Madison University and the University of Southern Mississippi.

Nobody seems to be asking what is wrong with selection for American and British universities when there are thousands of students who cannot do the academic work required while there are people without degrees who are able to produce acceptable work for relatively trivial wages. Some of the writers for the paper mills do have degrees. With their obvious and marketable research and writing skills shouldn't they be in doctoral programmes or academic appointments?

Is it possible that the trend towards holistic admissions in the US and the dumbing down of A-levels in the UK have something to do with it? If students are admitted to university on the basis of leadership, social skills displayed at interviews, participating in community service, overcoming adversity and the writing (by whom ?) of admission essays rather than the cognitive abilities and background knowledge necessary to do academic work then it would seem that the essay mill business is essential to keep the system going.

I have a suggestion for any university that wants to improve student and faculty quality within a short period. Find an essay mill operator, appoint him as admissions and recruitment officer and give scholarships to the essay writers who do not have degrees and faculty positions to those who do.

Sunday, March 08, 2009

Best Places for Postdocs

The Scientist has published its annual report on the best places for postdocs. A certain amount of scepticism is in order. There are dramatic changes in position from 2008 and 2009 -- for example, this year's number 1 in the US, the Whitehead Institute for Biomedical Research, was number 14 last year -- that suggest a certain amount of caution is needed.

Nonetheless, the results are interesting. First, in the US the tables are dominated by non-university institutions. Is it possible that politicisation and declining academic standards are beginning to have a noticeable effect on the quality of American university research?

Second, in the international category the top three positions are held by German, Danish and Dutch institutions. The only English university in the international top ten is York. What happened to Imperial College, Oxford and Cambridge?

Thursday, March 05, 2009

Interesting New Blog

Go here.


International Education Blogs ~ A New Blogging Project by David Comp
International Education Blogs is a new blogging project I started today as an effort to bring all of the blogs on the web that touch on international education issues into one central location. Essentially it will be a blog roll with a few postings. Please bear with me as I learn the technology. This initial launch will have two different blog lists for you to review.The first list of blogs will be those of people who are blogging on the field/state of international education and related matters.The second list of blogs will be those of students and others who are currently blogging from a foreign location.

Wednesday, February 25, 2009

Semantic Shift Watch

Inside Higher Ed has a story about Sonoma State University that starts


A faculty report has stirred some racial tensions at Sonoma State University, following claims from its author that the institution’s
administration has deliberately targeted those from higher-income families as potential students for the past decade. In this process, the report claims that the university has become the “whitest” public institution in California, effectively preferring white students to minorities in an admission practice that it deems “reverse affirmative action.”


One aspect of Sonoma State that is decidedly diverse is the administration, where the president, provost and director of admissions – all criticized in the report – are Latino. The
professor who brought forth this report, however, is white.


Since when is a university administration whose senior members are from one ethnic group decidedly diverse?

Friday, February 20, 2009

A Note on the THE-QS Employer Review

Of all the components of the THE-QS World University Rankings the employer review is probably the least noticed. Universities in South East Asia and elsewhere, for example, are going to great lengths to recruit international students to boost their performance in the rankings but there seems to be no comparable effort to do better in the employer review.

It might be worth looking at the structure of the review for 2008. First of all, the distribution by industry (2008) seems very unrepresentative with a disproportionate number of respondents drawn from financial services and banking, consulting and professional services, manufacturing and engineering and IT and computer services in that order.

The distribution by country (2006-8) is even more skewed. Take a look at the list of employer review resonses by country from the QS topuniversities site. The UK and Australia between them have more respondents than the USA. Mexico has more than China, Greece more than Germany, Singapore more than Japan, Ireland more than France and Romania more than any country in the Middle East, including Israel, Turkey, Iran and the Gulf states.

Incidentally, since the banking sector has been so incompetent at choosing whom to lend money to, is it a good idea to allow it to have such a big say in evaluating universities?




United States
346
United Kingdom
269
Australia
178
Mexico
75
Netherlands
75
Singapore
74
Russia
69
India
64
Argentina
60
Greece
59
Germany
56
Hong Kong
50
Philippines
45
Ireland
41
Malaysia
38
Canada
37
Japan
37
France
36
New Zealand
36
South Korea
32
Italy
29
Chile
28
Spain
27
Venezuela
27
China
25
Denmark
23
Thailand
23
Switzerland
22
Belgium
19
South Africa
19
Ukraine
18
Taiwan
17
Czech Republic
16
Romania
15
Other
354

Sunday, February 01, 2009

A Bit More About VU Amsterdam and the Previous Post

I have just remembered seeing this item at the bottom of the 2007 rankings on the QS site.

Vrije Universiteit AMSTERDAM

The data supplied by VU Amsterdam did not include faculty numbers for the VU Medical Center. Using 2006 citations data as a benchmark, it appears that the mapping of the citations database did not return the expected number of papers and citations - perhaps due to a
volume of research being published under institution names not easily
identifiable as being part of VU Amsterdam. The resolution of these issues would certainly result in a higher ranking for the university and improved performance in the Peer and Recruiter Reviews would suggest that VU Amsterdam may have maintained its Top 200 position.


It seems then that in 2007 QS did not count medical faculty but they did in 2008, resulting in a rise from 40 to 84 in the student- faculty ratio score. Secondly, in 2007 they did not count research if the affiliation did not clearly identify VU Amsterdam but they did in 2008, producing a rise in the citations score from1 to 38.

Saturday, January 31, 2009

No Way But Up

Researchtrends, a newsletter published by Scopus, has an interesting item on the THE-QS World University Rankings. It suggests that the rankings can be used to assess the recent research performance of countries as well as institutions.

This is not a bad idea in principle but researchtrends has failed to examine the rankings closely enough. Its writer observed that universities in some countries have improved their position quite dramatically. Two Indian institutions in the top 200 have enjoyed a net change in rank of 248 between 2007 and 2008, eleven from the Netherlands a change of 230 and seven Swiss universities one of 217 while three Israeli institutions rose 194 places between them. Researchtrends describe the Indian achievement as "astonishing" and "testament to the continued development of research in India. "

The writer concludes

"This suggests that national improvements in ranking may be at least
partially the result of individual universities taking a more strategic
approach: targeting international publications, aided by bibliometric tools and building and promoting library collections"

Are they?



Looking at the scores on the citations per faculty part of the rankings, we find that The Indian Institute of Technology (IIT) Bombay had exactly 1 point in 2007 and 43 in 2008, apparently a truly remarkable increase in research or rather citations of research.

The IIT Delhi went up from 1 to 47, the Free University of Amsterdam from 1 to 38, Technion Israel Institute of Technology from 1 to 79 and the Ecole Polytechnique Federale EPF) Lausanne from 29 to 77.

A couple of American universities also apparently enjoyed spectacular rises in the citations score, Washington University in St. Louis (WUSL), already discussed in this blog, and Stony Brook University from 1 to 75.

In reality, there has of course been no spectacular increase in research or citations. What has happened is that in 2007 QS had problems with identifying certain universities or got them confused with others. They, or THE, took a while to decide whether there was one IIT or several and almost certainly got confused between WUSL and the University of Washington. One reader of the blog has suggested they got the Free University of Amsterdam mixed up with the University of Amsterdam. Something similar could have happened with the EPF Lausanne.

So, there may have been an improvement in research productivity over the last few years in India, Switzerland, Israel and the Netherlands. But that is nothing to do with the ascent of some universities from these countries in the rankings. That is testament only to errors committed by QS or -- let us give them some credit -- the correction of errors.

Thursday, January 22, 2009

Top Universities Guide

I have finally forced myself to buy and read the second edition of QS's book, retitled Top Universities Guide.

The first edition, which was called Guide to the World's Top Universities, was an unqualified disaster, full of errors the worst of which was getting every single student faculty ratio wrong as a result of somebody moving every university three rows down while copying data. Others included putting the Technical University of Munich in the profiles twice in positions 82 and 98 and listing an "Official University of California, Riverside". It also started to fall apart within a few days.

The new edition, however, tuned out to be a pleasant surprise. The previous errors seem to have been corrected and there do not appear to be any new ones. There is a new dignified purple cover and so far not a single page has fallen out.

This does not mean that the THE-QS rankings are faultless. Far from it. The peer review remains extremely biased, the international components are largely meaningless and the student faculty ratio is misleading and easily manipulable. Still, the book does show that in a technical sense, QS has improved quite a bit recently.

The change in title should be noted. There may be another reason, but this would appear to be a clever attempt to distance QS and the authors from the first edition without admitting that there was anything wrong with it.

And I wonder why Blackwell is no longer distributing the book.

Sunday, December 21, 2008

A Must for University Ranking Fans

University rankings have been sprouting up all over the place recently. Laura Milligan at LearningXL has an annotated list of "100 Free College Rankings". It is very Americocentric -- neither the Shanghai nor the THE-QS rankings are there -- but there is something for nearly everybody. The strangest one is probably the Campus Squirrel Listings. The best, for those who are interested, include Berkeley, Mary Baldwin College and Kansas State University.

Friday, November 28, 2008

Comment on the THE-QS Academic Survey

QS have published some information on the academic survey that makes up 40 % of their World University Rankings. The table listing repondents by country is especially interesting. Here are some highlights

United States 638
United Kingdom 563
Australia 286
Inonesia 228
Malaysia 180
Philippones 201
France 125
Belgium 124
Germany 182
China 116
Kong Kong 100
Japan 96
Ireland 78
Thailand 43
Russia 41

So there are almost as many respondents from the UK as from the USA. There are more from Indonesia than from Germany. There are more from Thailand than from Russia. There are more from Hong Kong than from Japan.

Overall, there is pronounced bias in favour of the UK and Ireland, Southeast Asia and Australasia and against the USA, Japan and Southwest Asia.

Thursday, October 23, 2008

What does 100 on the academic survey mean?

This year 21 universities got a score of 100 on the academic survey section of the THE-QS rankings. A look at the subject rankings, which are combined with equal weighting to form the total score for this indicator, shows that such a perfect score can mean many different things.

Harvard scored 1oo in Arts and Humanities, 100 in Life Sciences and Biomedicine, 96.1 in Natural Sciences, 100 in Social Sciences and 59.6 in Engineering and IT.

The Australian National University scored 74 in Arts and Humanities, 46 .9 in Life Sciences and Biomedicine, 66.1 in Natural Sciences, 71.4 in Social sciences and 49.9 in Engineering and IT.

Peking University scored 56.4 for Arts and humanities, 56.9 for Life sciences and Biomedicine, 73 for Natural Sciences, 57.8 for Social Sciences, and 39.2 for Engineering and IT.

Friday, October 17, 2008

Farewell, Alabama

In 2007 the University of Alabama was listed by THE-QS as the fifth best research university in the world, as measured by citations per faculty. This year it is not even even in the top100. What happened?

Did all those researchers go on strike?

Or is just that that last year the number of faculty was underestimated and this year the mistake was corrected?

Tuesday, October 14, 2008

Ratio Inflation or Genuine Improvement?

QS have helpfully provided the means and standard deviations for their ranking indicators. For student faculty the mean is 0.09, which works out as a mean of 11.11 students per faculty.

But looking at the data for 2006, the mean number of students per faculty (using the scores provided by QS and cross-checking with the data on their website) was 16.44 students per faculty (N = 531). There has, it would seem, been a very substantial improvement on this indicator in only two years.

Is this a real improvement?

Saturday, October 11, 2008

Comment on a THE Editorial and Comments

Times Higher Education has published an editorial on the 2008 rankings that deserves comment. It says:

This is the fifth year we have published the rankings and the methodology
has remained unchanged for the past two. Along with our partners in the venture,
Quacquarelli Symonds, we make enormous efforts to ensure that our
quality-control processes and anti-cheating mechanisms are as robust as
possible.

I accept that so far this year the rankings have not been disfigured by the sort of spectacular errors that have occurred in the past but I would feel more confident about those enormous efforts if THE and QS owned up to their past errors, such as putting Duke in top place for student faculty ratio by counting undergraduate students as faculty, and indicated exactly what they are doing to stop similar mistakes from occurring again.

The editorial continues:

We try to ensure that the results are produced with a large amount of data. For example, there were more than 6,000 participants in the academic survey alone, producing an average of 20 responses per head. That is a staggering 120,000 data points, making it the largest known survey of university quality.

This is fallacious. The validity of any survey depends on how representative the respondents are of the larger population, not how many they are. If QS continued to report a response rate of around 3 percent they must expect continued criticism.

After proclaiming its lack of bias THE concludes:

For 2008, we congratulate Harvard University for its success
in topping the rankings yet again. However, it is worth remembering that its
endowment now totals more than $35 billion (£19 billion), roughly equivalent to
the total income received by the entire UK sector last year. By that measure,
the UK, with its 29 institutions in the Top 200 (and four in the Top 10), can
stand proud on the world stage.


According to THE's citations per faculty indicator the research impact of Cambridge academics is less than that of 35 other institutions including Resselaer Polytechnic Institute, Tufts, McMaster, Tel Aviv, UC Davis, Minnesota, Leiden, Emory, Toronto, Kyoto, Brown and ETH Zurch. Cambridge probably is not quite that bad -- there may be problems with the faculty side of the equation that are causing distortions -- but it does look as though the academic survey is in part an attempt to cover up the steady decline of British higher education and research.

There is a comment on the editorial by bgc:

The Times uses a non-transparent, undefined opinion
survey for most of their weightings - presumably this is what leads to such
nonsensical results.
Anyone who knows anything about international HE would
realize that the Times ranking lack basic validity.
For goodness sake,
sort-out the ranking methodology before next year. Or please stop inflicting
this annual embarrassment on those of us who practice scientometrics and try to
use objective methods of educational evaluation.



This is perhaps over-dramatic. The THE- QS rankings do seem to be improving in some respects. However, a reply by Martin Ince, editor of the rankings is rather unfortunate:

By contrast, we have measures relating to teaching,
globalisation and employability, and our research indicators cover the full
range of subjects. We set out exactly who and where our respondents are - there
is a nice pie chart in today's paper. These expert academics provide us with
about 126,000 data points (20 per person for 6,300 people) and make up the
biggest and best survey of university quality.



Whether the ability to sign on to a mailing list makes one an academic expert is debatable. And telling experts in scientometrics about your nice pie charts does you no good at all.

Friday, October 10, 2008

Not What It used To Be

The THE-QS rankings seem to be getting better. So far this year, no outrageous errors like ranking a non existent university (Beijing University), turning Malaysian ethnic minorities into foreigners or giving Washington University in St Louis a near zero for research, have surfaced.

But comparing the scores for the academic survey with those for citations for research, both of which are supposed to measure research quality, suggests that the former has is a large and systematic bias.

Here is a list of universities whose score on the academic survey exceeds their score on the citations per faculty by forty points or more.

Mcgill
Sydney
Melbourne
New York University
Monash
Trinity College Dublin
Peking
Seoul National University
Tsinghua
Auckland
London School of Economics
Warwick
Nanyang Technological University Singapore
University College Dublin
Fudan
Humboldt University Berlin
Nanjing
Shanghai Jiao Tong
Aotonomous National University of Mexico
Chulalongkorn University Thailand
Waseda
Lomonosov State University Moscow
Bologna
Sao Paulo
Beunos Aires

Maybe LSE and NYU can be explained by excellence in subjects that produce few publications or citations. But is it not possible that there is a pronounced geographical bias in the survey?

Thursday, October 09, 2008

THE-QS 2008 Update

This year there has been only one methodological change, namely the separation of the lists in the academic survey section into international and domestic sections and then their recombination. This would probably work against universities that receive a lot of votes from their own countries and might explain why Hong Kong, Peking and several Australian universities have fallen quite a bit.

Also, it is likely that the geographical spread of the academic and employer surveys has expanded and that this has benefitted universities in Latin America, Africa and India.

The biggest change in the top 100 is that Washington University in St Louis has risen to 60th place from 161st in 2007. This, presumably, is because it is now getting a realistic score for citations per faculty instead of the 1 it got in 2007, when QS seem to have confused it with the University of Washington. I am a little bit suspicious though about these two places being next to each other in this year's ranking.

The University of Hong Kong has fallen from 18th to 26th, Peking from 36th to 50th, Nanyang from 69th to 77th, Melbourne from 21st to 38th and Macquarie from 168th to 183rd (I wonder what Dr. Schwartz will say about that.)

On the other hand, the National Autonomous University of Mexico has risen from 192nd to 150th, the Indian Institute of Technology Delhi from 307th to 154th and Chulalongkorn from 223rd to 166th.

One oddity that I've noticed is that Stony Brook University, which is an autonomous university centre of the State University of new York, has risen dramatically to 127th place from 224th, while the other three centres at Binghamton, Buffalo and Albany which are of equal or better quality do not even get into QS's initial list.
The THE-QS Top 200: 2008

The QS topuniversities site has listed the top 200 universities in their 2008 rankings. I hope to make a few comments in a little while.