Field of Science

Showing posts with label Citizen Science. Show all posts
Showing posts with label Citizen Science. Show all posts

Citizen Science Works!

Earlier this month, we presented the results of the pilot phase of VerbCorner -- our citizen science project probing the nature of linguistic structure -- at a scientific conference (the Workshop on Events in Language and Cognition). You can see the poster describing the work here

For those who don't know or don't remember, in VerbCorner, we're trying to work out the grammar rules that apply to verbs. Why do you say Agnes looked at the wall but not Agnes saw at the wall? Why do you say Bart filled the glass with water but not Bart poured the glass with water? Many -- but not all -- linguists believe that grammatical idiosyncrasies are explained by the meanings of the verbs, but evidence is sketchy. Volunteers have been visiting our website to help analyze the meanings of verbs so we can find out.

High-Quality Analyses by Volunteers

Our initial work -- the pilot for the pilot, if you will -- suggested that we could get high-quality analyses from volunteers. But that was based on a very small sample. As of late Feb, over 10,000 volunteers had contributed over 525,000 analyses. In general, the agreement between different volunteers was pretty high -- which is a good sign. Just as importantly, we had a smaller set of 'test' items, for which we knew what professional linguists would say. When we combine the analyses of different volunteers for the same sentence in order to get a 'final answer', the results match the analyses of professional linguists very well. This shows that we can trust these results.

Where Quantity Becomes Quality

Just as importantly, we were able to analyze a lot of sentences. In the VerbCorner project, we are trying to determine which sentences have which of a very specific set of aspects of meaning. One aspect is whether the sentence involves something changing physical form (example: Agnes broke the vase as opposed to Agnes touched the vase). Another aspect is whether the sentence involves anything applying physical force to anything else (ex: Agnes pushed Bart as opposed to Agnes looked at Bart). 

For purposes of bookkeeping, let's call one aspect of meaning for one sentence an 'item.' After combining across different volunteers, the results were clear enough to definitively code 31,429 items. This makes VerbCorner the largest study of it's kind by far. (A typical study might only look at a few hundred items.) 

This quantity makes a big difference. Given how small studies usually are, they can only look at one tiny corner of the language. The problem is that that corner might not be representative. Imagine studying what Americans are like by only surveying people in Brooklyn. This tends to lead to disagreements between different studies; one linguist studies "Brooklyn" and another studies "Omaha", and they come to very different conclusions! Unfortunately, language is so complex and so vast, one person can only analyze one corner. This is why we are recruiting large numbers of volunteers to help!

The results

One major question we had was how much the rules of verb argument structure (that is, the kinds of grammatical rules described above) depend on meaning. Some linguists think they depend entirely on meaning: If you know the meaning of a verb, you know what its grammar will be like. Others think meaning has very little role to play. Most linguists are probably somewhere in the middle.

The results suggest that the first group is right: These rules depend almost entirely on meaning. Or maybe even entirely; it's so close it is hard to tell.

The reason I say "suggest," however, is that while we have the biggest study of its kind, it still only covers about 1% of English. So we've gone from studying Brooklyn to studying all of NYC. It's an improvement, but not yet enough. 

This is why I called this first phase a "pilot". We wanted to see if we could get high-quality, clearly-interpretable results from working with volunteers. Many researchers thought this would be impossible. After all, linguists have to go through a lot of schooling to learn how to analyze sentences. But a key finding of the Citizen Science movement is that there are a lot of smart enthusiasts out there who may not be professionals but can very much contribute to science.

The next phase

We have set a goal of reaching 50,000 completed items by July 1st. That will require upping our game and increasing the rate at which we're analyzing items by almost 4x. But the beauty of Citizen Science is that this does not really require that much work on anyone's part. If 3,000 volunteers each spend about one hour contributing to the project, we'll more than hit that goal. So please help out, and please tell your friends. You can contribute here.

More Citizens, More Science


For the last couple years, most articles about Citizen Science -- in which amateurs contribute to scientific projects -- have been hagiography. These articles were nearly exclusively Ra! Ra!, all about the exciting new development.

It seems that we've matured a bit as a field, because lately I've run across a couple articles that, while still being positive overall, have laid out some real criticism. For instance, in an article in Harvard Magazine, Katherine Xue concludes with the worry that citizen science may be less about involving the public and more about cheap labor (full disclosure: I was interviewed for and appear in this article). Many citizen science projects, she notes, are little more than games or, worse, rote labor, with little true engagement for the volunteer in the scientific mission.

Similarly, in a much-tweeted article at The Guardian, Michelle Kilfoyle and Hayley Birch write, "Who really benefits the most from [citizen science]: the amateurs or the professionals? … Most well-known initiatives are the big crowdsourcing projects: big on the number of participants but not necessarily the level of participation."

Introducing the VerbCorner Forum

These articles resonated with me. Ever since we launched VerbCorner, our citizen science project looking at the structure of language, meaning, and thought, we've wanted to find additional ways to get our volunteers involved in the science and get more out of participation. VerbCorner is very much a crowdsourcing project -- most of what volunteers do on the site is contribute labor. We've always had this blog, where people could learn more about the project, but that's not especially interactive.

To that end, we've added a forum where anyone and everyone involved in the project can discuss the project, offer suggestions, debate the science, and discuss anything related (syntax, semantics, etc.). We have high hopes for this forum. Over the years, I have gotten a lot of emails from participants in the various projects at GamesWithWords.org, emails with questions about the projects, ideas for new experiments, and -- all too often -- reports of bugs or type-os. These emails have been extremely useful, and in a few cases have even led to entirely new research directions. But email is a blunt instrument, and I expect that for everyone who has emailed, at least ten others had similar comments but never got around to tracking down our email address.

I hope to see you on the forum!

GamesWithWords on Scientific American

Over the last week, ScientificAmerican.com has published two articles by me. The most recent, "Citizen Scientists decode meaning, memory and laughter," discusses how citizen science projects -- science projects involving collaborations between professional scientists and amateur volunteers -- are now being used to answer questions about the human mind.

Citizen Science – projects which involve collaboration between professional scientists and teams of enthusiastic amateurs — is big these days. It’s been great for layfolk interested in science, who can now not just read about science but participate in it. It has been great for scientists, with numerous mega-successes like Zooniverse and Foldit. Citizen Science has also been a boon for science writing, since readers can literally engage with the story.
However, the Citizen Science bonanza has not contributed to all scientific disciplines equally, with many projects in zoology and astronomy but less in physics and the science of the mind. It is maybe no surprise that there have been few Citizen Science projects in particle physics (not many people have accelerators in their back yards!), but the fact that there has been very little Citizen Science of the mind is perhaps more remarkable.

The article goes on to discuss three new mind-related citizen science projects, including our own VerbCorner project.

The second, "How to understand the deep structures of language," describes some really exciting work on how to explain linguistic universals -- work that was conducted by colleagues of mine at MIT.
In an exciting recent paper, Ted Gibson and colleagues provide evidence for a design-constraint explanation of a well-known bias involving case endings and word order. Case-markers are special affixes stuck onto nouns that specify whether the noun is the subject or object (etc.) of the verb. In English, you can see this on pronouns (compare "she talked with her"), but otherwise, English, like most SVO languages (languages where the typical word order is Subject, Verb, Object) does not mark case. In contrast, Japanese, like most SOV languages (languages where the typical word order is Subject, Object, Verb) does mark case, with -wa added to subjects and -o added to direct objects. "Yasu saw the bird" is translated as "Yasu-wa tori-o mita" and "The bird saw Yasu" is translated as "Tori-wa Yasu-o mita." The question is why there is this relationship between case-marking and SOV word order.
The article ran in the Mind Matters column, which invites scientists to write about the paper that came out in the last year that they are most excited about. It was very easy for me to choose this one.

VerbCorner (and others) on SciStarter.Com

There is a brief profile of our crowd-sourcing project VerbCorner on SciStarter.com, with a number of quotes form yours truly.

SciStarter profiles a lot of Citizen Science / Crowd-sourced Science projects. Interestingly, most are physical sciences, with only one project listed under psychology (interestingly, also a language project).

This is not a feature of SciStarter but more a feature of Citizen Science. The Scientific American database only lists two projects under "mind and brain" -- and I'm pretty sure they didn't even have that category last time I checked. This is interesting, because psychologists have been using the Internet to do research for a very long time -- probably longer than anyone else. But we've been very late to the Citizen Science party.

Not, of course, that you shouldn't want to participant in non-cognitive science projects. There are a bunch of great ones. I've personally mostly only done the ones at Zooniverse, but SciStarter lists hundreds.