Altmetrics: what’s the point?

A couple of weeks ago Stephen (of this parish) generated a lot of discussion when he complained about the journal impact factor (JIF). I must admit I feel a bit sorry for the JIF. It’s certainly not perfect, but it’s clear that a lot of problems aren’t with the statistic itself, but rather with the way it is used. Basically, people take it too seriously.

The JIF is used as a measure of quality of science: it’s use to assess people and departments, not journals. And this can affect funding decisions, and hence people’s careers. If we are to use a numerical metric to make judgements about the quality of science being done, then we need to be sure that it is actually measuring quality. The complaint is that the JIF doesn’t do a good job of this.

But a strong argument can be made that we do need some sort of measure of quality. There are times when we can’t judge a person or a department by reading all of their papers: a few years ago I applied for a job for which there was over 600 applicants. Imagine trying to read 3 or 4 papers from each applicant to get an idea about how good they are.

Enter the altmetrics community. They are arguing that they can replace the JIF with better, alternative metrics. They are trying to develop these metrics using new sources of information that can be used to measure scientific worth: online (and hence easily available) sources like twitter and facebook (OK, and also Mendeley and Zotero, which make more sense).

Now, I have a few concerns about altmetrics: they seems to be concentrating on using data that is easily accessible and which can be accumulated quickly, which suggests that they are interested in work which is quickly recognised as important. Ironically, one of the criticisms of the JIF is that it only has 2 year window, so down-grades subjects (like the ones I work in) which have a longer attention span.

But I also have a deeper concern, and one I haven’t seen discussed. It’s a problem that, if it is not solved, utterly undermines the altmetrics programme. It’s that we have no concrete idea what it is they are trying to measure.

The problem is that we want our metrics to capture some essence of the influence/impact/importance that a paper has on science, and also on the wider world. But what do we mean by “influence”? It a very vague concept, so how can we operationalise the concept? The JIF at does this by assuming that influence = number of citations. This has some logic, although it limits the concept of influence a lot. It also assumes that all citations are equal, irrespective of the reason for citation or where the citing paper is published. But in reality these things probably matter: being cited in an important paper is worth more than in a crappy paper that nobody is going to read.

But what about comparing a paper that has been cited once in a Very Important Paper to one that has been cited three times in more trivial works. Which one is more important? In other words, how do we weight number of citations against importance of where they are cited to measure influence?

I can’t see how we can even start to do this if we don’t have any operational definition of influence. Without that, how can we know whether any weighting is correct? Sure, we can produce some summary statistics, but if we don’t even know what we’re trying to measure, how can we begin to assess if we’re measuring it well?

I’ve sketched the problem in the context of citations, but it gets even worse when we look at altmetrics. How do we compare tweets, Facebook likes, Mendeley uploads etc? Are 20 tweets the same as one Mendeley upload? Again, how can we tell if we can’t even explicate what we are measuring?

If someone can explain how to do this, then great. But I’m sceptical that it’s even possible: I can’t see how to start. And if it isn’t possible, then what’s the point of developing altmetrics? Shouldn’t we just ditch all metrics and get on with judging scientific output more qualitatively?

Unfortunately, that’s not a realistic option: as I pointed out above, with the amount of science being done, we have to use some form of numerical summary, i.e. some sort of metric. So we’re stuck with the JIF or other metrics, and we can’t even decide if they’re any good.

Bugger.

Posted in The Society of Science, Uncategorized | 41 Comments

A new really old version of The Elements

I hope you’re all fans out Tom Lehrer (“Mr. Lehrer’s muse [is] not fettered by such inhibiting factors as taste.” – NYT, apparently). Well, via those Improbable Research (“Mr. Abraham’s muse [is] not fettered by such inhibiting factors as taste.” – NYT, possibly) comes a new old version of Lehrer’s classic The Elements:

The lyrics are below the fold.
Continue reading

Posted in Uncategorized | 3 Comments

Research with impact

After Stephen’s posts about impact factors and the like, I have a couple of serious posts brewing. But for now (and because it’s Friday), I want to admit to my reaction today to an advert I got about a journal, which told me that I should Stand Out in my Field, and Be Visible by submitting to this journal (run by a reputable publisher). One of the reasons for publishing there was that the journal has High Impact: it’s impact factor is 1.95.

“Pah!” I thought. I’m executive editor of a journal with an impact factor over 5. In fact 5.093 (that 3 at the end is terribly important, no matter what anyone tells you). Why would I want to publish is such a lowly journal?

On the other hand, the journal has published a paper about TARDIGRADES IN SPACE. That beats impact factors any day.

Posted in Science Publishing, Silliness | 9 Comments

lme4: destined to become stable through rounding?

(this would have appeared on my blog on Nature Network, but the pulled the plug the day before. Sometimes correlation does not mean causation)

Fans of R and mixed models are aware of the lme4 package. This started out as Doug Bates re-writing the lme package using the new capabilities in R (S4 objects, for those who care about such things). It goes back to at least 2006, but isn’t stable yet: a source of mild amusement for me over the last few years. In software development, an un-stable version has a number starting with 0 (e.g. 0.4), and once the developers are happy with it, it gets upgraded to v1.0. The core R developers released R1.0-0 on the 29th February 2000, citing it as the nerdiest date possible, being an exception to an exception.
Anyway, the version numbers of lme4 show the problems: v0.999375 was released in 2008. I just checked and the latest version is 0.999999-0-1. This is more compactly (and more confusingly) written as 1-1e06-0-1.
I have been worried that lme4 will never become stable, but this latest version mollifies me with the thought that the developers can’t go on forever, so eventually lme4 will become stable when the machine precision forces it to be rounded up to 1.0.

Posted in R, Silliness | Comments Off on lme4: destined to become stable through rounding?

Welcome to The Menagerie

In my first OT post I mentioned The Menagerie I live in. So, while GrrlScientist is attending to parts of it I thought I’d introduce some of the residents, including some of the shyer ones

ShrimpLaceleaf
Continue reading

Posted in Life in the Menagerie | 12 Comments

Stupidity Molecule Identified

Researchers at the University of Utrecht have identified a molecule that could play a key role in controlling how stupid headlines develop in scientific press releases.

Neurobiologists are trying to understand how titles that are at first sensible differentiate into specialised pieces of text that are sensationalist and which can be classified into two groups, by being scientifically misleading or jibberish. Researchers in the College of Science Sciences at Utrecht, led by Professor H.G.P. Strabismus, have identified a molecule called methyl-methanol as being the ‘signal’ which can induce the production of these headlines.

The Strabismus laboratory studies a simple system, the Press Officer, which processes complex scientific results and reduces them to their bare essentials, which are then presented to the public. In earlier research they showed that methyl-methanol induces the introduction of errors into these presentations. Now they have identified a more specialised role for the molecule, specifically targetting the headlines.

The new research will be published in the journal Nature, once the brown envelope has arrived.

“Our work presents the opportunity to fully understand how POs learned to utterly mangle explanations of science. The actual work was looking at cellular differentiation in prokaryotessaid Professor Strabismus.

“These findings are also remarkable because methyl-methanol was previously found to affect celebrities, causign them to lose motility and transform into large sticky messes, known as headline fodder. The fact that an individual like a Press Officer, which is very far removed from celebrity, uses the same mechanism is very interesting and suggests that the processes which cause cock-ups may have very deep evolutionary origins.”

The work has been funded by the Unwellcome Trust and CAMRA.

ENDS

(hat tip: a BBSRC press officer for the inspiration)

Posted in Silliness | 7 Comments

Helloooo! I’ve moved (again)

I started blogging just over 5 years ago (damn, I missed my blogoversary by a month). The following year, I left Blogspot for Nature Network. Well, now the powers that be have decided to close the NN blogs, and shift bloggers who wanted over to SciLogs. I considered this, but decided instead to join the hive of villeins here at Occam’s Typewriter.

My NN archives should me moved across here soon, once the virtual workmen have finished their virtual cups of tea. I’ll also add a blogroll and whatever else takes my fancy. I’ll also try to write some new blogposts.

For those of you who don’t know what to expect, it’s OK, I’m not sure either. The name of the blog was always going to be descriptive, so it’ll either be something silly (and probably fairly short), or a more considered (=boring) posts about statistics, biology, science, life etc.

Professionally I’m a statistician in the Biodiversity and Climate Change Research Centre, a.k.a. BiK-F. Most of my work surrounds the torture and confession of ecological and evolutionary data. I’m also now executive editor of the very fine journal Methods in Ecology and Evolution, so you might see some blog posts from me pop up there too. Oh, and I occasionally sneak a post onto my wife’s Guardian blog, to dilute the mystery birds.

Most importantly, the blog’s banner is a photo of The Beast, the Aussie cat who I serve. He lives with GrrlScientist and I, and the rest of our menagerie (3 parrots, N passerines, and the contents of 5 aquaria) The Beast’s main role is to provide hair for recycling. If anyone knows what to do with the stuff, I’d love to know.

That’s me, and I’ll be here for a bit (hopefully). Now, who are you? Please, speak up so I know if anyone’s reading this stuff.

Posted in Uncategorized | 20 Comments

Abusing a Prior: some slides

Here are the slides for my talk today about Bayesian variable selection. It’s mainly of interest to other statisticians, my excuse is that I’m talking at a statistical meeting (and I’m a keynote speaker! Wooo!).

Continue reading

Posted in Statistics | Comments Off on Abusing a Prior: some slides

Fossils, Fun, and Frailities. And Statistics #isec2012

If I’ve done everything correctly, then at almost the moment this post appears, I will start talking about some fossils. Not, it’s not a eulogy about defenders of the 4-4-2 formation, but it’s about a bit of work I’ve been doing on trying to model when fossils appear and disappear. The work is still preliminary, in some ways, and it would be nice to get some feedback. So here are the slides:

Continue reading

Posted in Research Blogging, Statistics | Comments Off on Fossils, Fun, and Frailities. And Statistics #isec2012

Today’s quiz: explain the obscure site

Can anyone tell me what is probably so famous about this site, in Essex? The bit of historical interest is next to the houseboat in the centre of the image.
View Larger Map
If you’re on twitter, or have read the right blogs, you know already. So no giving it away. And googling it is cheating. 🙂
And bonus points will be given for explaining the rectangles.

Posted in Science Blogging | 6 Comments