Field of Science

Talking about Love

Much of my work is on verbs that describe emotion, called "psych verbs." The curious thing about psych verbs is that they come in two varieties, those that put the experiencer of the emotion in subject position (Mary likes/hates/fears John) and those that put the experiencer of the emotion in object position (Mary delights/angers/frightens John).

These verbs have caused a four-decades-long headache for theorists trying to explain how people know what should be the subject and what should be the object of a given verb. Many theorists would like to posit theories on which you put the "do-er" in subject position and the one "done to" in object position. But some psych verbs seem to go one way and some the other.

There are basically only three theoretical possibilities:

a) There's no general rule that will tell you whether the experiencer of an emotion should be the subject or object of a given verb.

b) There's a general rule that tells you the experiencer should be the subject (or, on other theories, the object), and then there are some exceptions.

c) There are no exceptions. There are two kinds of psych verbs that actually mean very different things. Each group follows a particular rule: one sends the experiencer to subject; the other, to object.

I started out as a fan of theory (b). The results of my own work have pushed me in the direction of (c). The only theory that I'm pretty sure is wrong is (a). There are a lot of reasons I think (a) is wrong. One has to do with Broca's aphasia.

Broca's aphasia

People with Broca's aphasia -- typically caused by a stroke or brain injury -- have difficulty with grammar but are relatively good at remembering what individual words mean. Classically, Broca's aphasia was thought to result from damage to Broca's area, though I've heard that association is not as solid as once believed.
Some well-known language-related areas of the brain.

Either way, Maria Mercedes Pinango published a study in 2000 looking at how well Broca's aphasics understand psych verbs. She found that they had particular trouble with experiencer-object verbs (delights/angers/frightens) ... unless the verbs were in passive form (Mary is delighted/angered/frightened by John), in which case they had more trouble with the experiencer-subject verbs.

There are a lot of reasons this could be. The main aspect of the finding that interests me here is that this is *not* what you'd expect on theory (a), since on that theory, all psych verbs are more or less the same and there's no particular reason Broca's aphasia or anything else should impact one more than the other.

One worry one might have about this study was that it was published as a book chapter and not in a journal, and book chapters don't (usually) undergo the same review process. I don't personally know that much about aphasia or how one goes about testing aphasics, so it's hard for me to review Pinango's methods. More importantly, there weren't many participants in the study (these participants are not easy to find), so one would like replication.

Replication


As it happens, Cynthia Thompson and Miseon Lee recently published just such a replication (well, they published it in 2009, but one doesn't always hear about papers right away). It's a nice study with 5 Broca's aphasics, published in the Journal of Neurolinguistics. They tested both sentence comprehension and sentence production, finding that while passive sentences were harder overall, experiencer-subject verbs (like/hate/fear) were easier in the active form and experiencer-object verbs (delight/anger/frighten) were easier in the passive form. This effect was much more pronounced in sentence production than comprehension (in the latter case, it was not strictly significant), most likely because comprehension is easier.

Again, these are not the results you expect if the rules that tell you who should be a subject and who should be an object are verb-by-verb, since then there's no reason brain damage should affect one class of verbs as opposed to another (since there are no verb classes).* What exactly it does mean is much trickier. Give me another 20-30 years, and hopefully I'll have an answer.





*Actually, I can come up with a just-so story that saves theory (a). But it's certainly not what you would expect, and I believe there are a lot of other data from other paradigms that speak against theory (a).

_________

Thompson CK, and Lee M (2009). Psych verb production and comprehension in agrammatic Broca's aphasia. Journal of neurolinguistics, 22 (4), 354-369 PMID: 20174592

New York Times, You Can't Handle the Truth.

Earlier today I wrote about the research behind an opinion article at the New York Times. When I looked at the sources cited, I was unable to find any information supporting the claims made in the article. In fact, what I found directly contradicted those claims. I finished by saying that while I was willing to believe these claims, I'd like to know what data support them. In passing, I mentioned that I had submitted an abbreviated version of this analysis as a comment on the Times website.

That comment was not published. I figured maybe there had been a computer error, so I submitted another one later in the day. That one was also not published. Finally, at 6:13pm, I submitted an innocuous and useless comment under an assumed name:
I agree with Pat N. It's nice to hear from someone who has some optimism (@ Dr. Q).
This comment was published almost immediately.


The Times states that "comments are moderated and generally will be posted if they are on-topic and not abusive."Since the moderators didn't publish the comment, we can conclude one of two things:

1) Discussion of the empirical claims made in a New York Times article is not "on topic."
2) Pointing out a mistake made in a New York Times article is a kind of abuse.

Do students at selective schools really study less?

*Updated with More Analysis*


So says Philip Babcock in today's New York Times. He claims:
Full-time college students in the 1960s studies 24 hours per week, on average, and their counterparts today study 14 hours per week. The 10-hour decline is visible for students from all demographic groups and of all cognitive abilities, in every major and at every type of college.
The claim that this is true for "every type of college" is important because he wants to conclude that schools have lowered their standards. The alternative is that there are more, low-quality schools now, or that some schools have massively lowered their standards. These are both potentially problems -- and are probably real -- but are not quite the same problem as all schools everywhere lowering their standards.

So it's important to show that individual schools have lowered their standards, and that this is true for the selective schools as well as the not-selective schools. The article links to this study by Babcock. This study analyzes a series of surveys of student study habits from the 1960s to the 2000s, and thus seems to be the basis of his argument, and in fact the introduction contains almost the identical statement that I have quoted above. Nonetheless, despite these strong conclusions, the data that would support them appear to be missing.
SAT scores and size are not available in the early years, so study time by college selectivity is not reported. 
He goes on to say that he can look at selectivity in the more recent surveys: specifically matched 1988-2003 surveys. These do show a decrease in study time from on the order of 1-2 hours for high-, medium- and low-selectivity schools (I cannot find how selectivity was defined). Whether this is even statistically significant is unclear, as he does not report any statistics or confidence intervals. In any case, it is not a 10 hour difference.

What Babcock might have meant, and more problems with the data

It is possible that when Babcock was saying that the decrease in study time was true of all types of schools, he meant that when you look at all types of schools in 2003/4, students at all levels report studying less than the average student reported in 1961. The problem is that, for all we know, the schools in his sample were more selective in 1961 than they were in 2003/4.

Moreover, the is something worrisome about his selectivity data. Whenever analyzing data, many researchers like to do what is called a "sanity check": they make sure that the data contain results that are known to be true. If you were looking at a study of different types of athletes, you might make sure that the jockeys are shorter than the basketball players, lighter than the football players and chew less tobacco than the baseball players. If you find any of these things do not hold, you might go back and make sure there isn't a type-o somewhere in your data-entry process.

I worry that Babcock's data fail the sanity check. Specifically, look at the number of hours studies according to selectivity of school in 2003:

highly selective: 13.47 hours
middle:               14.68 hours
non-selective:     16.49 hours

Note that this effect is larger than the decline in number of hours studied between 1988 and 2003, so in terms of this dataset, this is a large effect (again, I cannot tell if it is significant, because the relevant statistical information is not provided) and it's not in the direction one would think. I will admit that it is possible that students at highly selective schools really do study less than the folks at JuCo, but that conflicts heavily with my pretty extensive anecdotal database. So either a) the world is very different from how I thought it was -- in which case, I want more evidence than just this survey -- b) Babcock has defined selectivity incorrectly, or c) there is something wrong with these data.

One last worrisome fact

I considered the possibility that the data Babcock was quoting were in a different paper. The only other paper on Babcock's website that looked promising was this American Enterprise Institute report. This is not a research paper, but rather summarizes research. Specifically, according to footnote #2, it summarizes the research in the paper I just discussed. Strangely, this paper does have a graph (Figure 4) breaking down study habits of students in the 1960s based on selectivity of the school they are attending: the very data he states do not exist in the later paper.

I'm not really sure what to make of that, and have nothing further to say on the topic. At the very least, I would be hesitant to use those graphs as evidence to support the general claim that study habits have changed even at the selective schools, since it's unclear where the data case from, or if in fact they even exist (to be clear: it's Babcock who says they don't exist, not me).

Conclusion

To summarize, there seems to be very little evidence to support Babcock's conclusion that study time has decreased even at selective schools by 10 hours from the 1960s to modern day. That is, he has a survey from 1961 in which students studied 25 hrs/week, two surveys in the 1980s in which students studied 17 hours/week, and two surveys in the 2000s in which students studied 14-15 hrs/week, but these surveys are all based on different types of students at different schools, so it's hard to make any strong conclusions. If I compared the weight of football places from Oberlin in 1930 and Ohio State in 2005, I'd find a great increase in weight, but in fact the weight of football players at Oberlin probably has not increased much over that time period.

Moreover, there are aspects of these data that deserve some skepticism. When report to people who went to selective schools that these data suggest students at such schools study 13 hrs/week, the response is usually something like, "Do you mean per day?"

Finally, since no statistics were run, it's quite possible that none of the results in this study are significant.

I want to be clear that I'm not saying that Babcock's claims aren't true. I'm just saying that it's not clear he has any evidence to support them (which is not to say I think it's a bad study: it was a good study to have done and clearly took a lot of work, but I find it at best suggestive of future avenues of research and certainly not conclusive).

New tags

Rather than write a new blog post (or my nearly-due BUCLD proceedings paper), I decided to revamp the post tags on this blog. Their usage has been inconsistent, which is making it harder and harder to find old blog posts that I want to link to.

Hopefully the new and improved tags will also be useful for you, dear reader. Now if you want to find any of my articles on the academic career path, on animal language or on universal grammar -- just to give a few examples -- they are only a mouse click away.

In addition to standard tags, there are also a series of tags beginning with the preposition "on". These appear on most posts now and are more meta-thematic than the others.

Learning What Not to Say

A troubling fact about language is that words can be used in more than one way. For instance, I can throw a ball, I can throw a party, and I can throw a party that is also a ball.

These cats are having a ball.

The Causative Alternation

Sometimes the relationship between different uses of a word is completely arbitrary. If there's any relationship between the different meanings of ball, most people don't know it. But sometimes there are straightforward, predictable relationships. For instance, consider:

John broke the vase.
The vase broke.

Mary rolled the ball.
The ball rolled.

This is the famous causative alternation. Some verbs can be used with only a subject (The vase broke. The ball rolled) or with a subject and an object (John broke the vase. Mary rolled the ball). The relationship is highly systematic. When there is both a subject and an object, the subject has done something that changed the object. When there is only a subject, it is the subject that undergoes the change. Not all verbs work this way:

Sally ate some soup.
Some soup ate.

Notice that Some soup ate doesn't mean that some soup was eaten, but rather has to mean nonsensically that it was the soup doing the eating. Some verbs simply have no meaning at all without an object:

Bill threw the ball.
*The ball threw.

In this case, The ball threw doesn't appear to mean anything, nonsensical or otherwise (signified by the *). Try:

*John laughed Bill.
Bill laughed.

Here, laughed can only appear with a subject and no object.

The dative alternation

Another famous alternation is the dative alternation:

John gave a book to Mary.
John gave Mary a book.

Mary rolled the ball to John.
Mary rolled John the ball.

Once again, not all verbs allow this alternation:

John donated a book to the library.
*John donated the library a book.

(Some people actually think John donated the library a book sounds OK. That's all right. There is dialectical variation. But for everyone there are verbs that won't alternate.)

The developmental problem


These alternations present a problem for theory: how do children learn which verbs can be used in which forms? A kid who learns that all verbs that appear with both subjects and objects can appear with only subjects is going to sound funny. But so is the kid who thinks verbs can only take one form.
The trick is learning what not to say

One naive theory is that kids are very conservative. They only use verbs in constructions that they've heard. So until they hear "The vase broke," they don't think that break can appear in that construction. The problem with this theory is that lots of verbs are so rare that it's possible that (a) the verb can be used in both constructions, but (b) you'll never hear it used in both.

Another possibility is that kids are wildly optimistic about verb alternations and assume any verb can appear in any form unless told otherwise. There are two problems with this. The first is that kids are rarely corrected when they say something wrong. But perhaps you could just assume that, after a certain amount of time, if you haven't heard e.g. The ball threw then threw can't be used without an object. The problem with that is, again, that some verbs are so rare that you'll only hear them a few times in your life. By the time you've heard that verb enough to know for sure it doesn't appear in a particular construction, you'll be dead.

The verb class hypothesis

In the late 1980s, building on previous work, Steven Pinker suggested a solution to this problem. Essentially, there are certain types of verbs which, in theory, could participate in a given alternation. Verbs involving caused changes (break, eat, laugh) in theory can participate in the causative alternation, and verbs involving transfer of possession (roll, donate) in theory can participate in the dative alternation, and this knowledge is probably innate. What a child has to learn is which verbs do participate in the dative alternation.

For reasons described above, this can't be done one verb at a time. And this is where the exciting part of the theory comes in. Pinker (building very heavily on work by Ray Jackendoff and others) argues that verbs have core aspects of their meaning and some extra stuff. For instance, break, crack, crash, rend, shatter, smash, splinter and tear all describe something being caused to fall to pieces. What varies between the verbs is the exact manner in which this happens. Jackendoff and others argues that the shared meaning is what is important to grammar, whereas the manner of falling to pieces was extra information which, while important, is not grammatically central.

Pinker's hypothesis was that verb alternations make use of this core meaning, not the "extra" meaning. From the perspective of the alternation, then, break, crack, crash, rend, shatter, smash, splinter and tear are all the same verb. So children are not learning whether break alternates, they learn whether the whole class of verbs alternate. Since there are many fewer classes than than there are verbs (my favorite compendium VerbNet has only about 270), the fact that some verbs are very rare isn't that important. If you know what class it belongs to, as long as the class itself is common enough, you're golden.

Testing the theory


This particular theory has not been tested as much as one might expect, partly because it is hard to test. It is rather trivial to show that verbs do or don't participate in alternations as a class, partly because that's how verb classes are often defined (that's how VerbNet does it). Moreover, various folks (like Stefanowitsch, 2008) argue that although speakers might notice the verb classes, that doesn't prove that people actually do use those verb classes to learn which verbs alternate and which do not.

The best test, then, is it teach people -- particularly young children -- new verbs that either belong to a class that does alternate or to a class that does not and see if they think those new verbs should or should not alternate. Very few such studies have been done.

Around the same time Pinker's seminal Language and Cognition came out in 1989, which outlines the theory I described above, a research team led by his student Jess Gropen (Gropen, Pinker, Hollander, Golberg and Wilson, 1989) published a study of the dative alternation. They taught children new verbs of transfer (such as "moop," which meant to move an object to someone using a scoop), which in theory could undergo the dative alternation. The question they asked was whether kids would be more likely to use those verbs in the alternation if the verbs were monosyllabic (moop) or bisyllabic (orgulate). They were more likely to do so for the monosyllabic verbs, and in fact in English monosyllabic verbs are more likely to alternate. This issue of how many syllables the verb has did come up in Language and Cognition, but it wasn't -- at least to me -- the most compelling part of the story (which is why I left it out of the discussion so far!).

Ambridge, Pine and Rowland (2011)


Ben Ambridge, Julian Pine and Caroline Rowland of the University of Liverpool have a new study in press which is the only study to have directly tested whether verb meaning really does guide which constructions a child thinks a given verb can be used in, at least to the best of my knowledge -- and apparently to theirs, since they don't cite anyone else. (I've since learned that Brooks and Tomasello, 1999, might be relevant, but the details are sufficiently complicated and the paper sufficiently long that I'm not yet sure.)

They taught children two novel verbs, one of which should belong to a verb class that participates in the causative alternation (a manner of motion verb: bounce, move, twist, rotate, float) and one of which should not (an emotional expression: smile, laugh, giggle). Just to prove to you that these classes exist, compare:

John bounced/moved/twisted/rotated/floated the ball.

The ball bounced/moved/twisted/rotated/floated.


*John smiled/laughed/giggled Sally.
Sally smiled/laughed/giggled.

Two groups of children (5-6 years old and 9-10 years old) were taught both types of verbs with subjects only. After a lot of training, they were shown new sentences with the verbs and asked to rate how good the sentences were. In the case of the manner of motion verb, they liked the sentences whether the verb had an subject and an object or if the verb had only a subject. That is, they thought the verb participated in the causative alternation. For the emotion expression verb, however, they thought it sounded good with a subject only; when it had both a subject and an object, they thought it did not sound good. This was true both for the older kids and the younger kids.

This is, I think, a pretty nice confirmation of Pinker's theory. Interestingly, Ambridge and colleagues think that Pinker is nonetheless wrong, but based on other considerations. Partly, our difference of opinion comes from the fact that we interpret Pinker's theory differently. I think I'm right, but that's a topic for another post. Also, there is some disagreement about a related phenomenon (entrenchment), but that, too, is a long post, and the present post is long enough.


____
Gropen, J., Pinker, S., Hollander, M., Goldberg, R., and Wilson, R. (1989). The Learnability and Acquisition of the Dative Alternation in English Language, 65 (2) DOI: 10.2307/415332


Ben Ambridge, Julian M. Pine, and Caroline F. Rowland (2011). Children use verb semantics to retreat from overgeneralization errors Cognitive Linguistics


For picture credits, look here and here.

New Experiment: EmotionSense

I just posted a new experiment on the website: EmotionSense. I have lately gotten very interested in verb-learning, specifically how we decide which of the participants in an event should be the grammatical subject, which the grammatical object, etc. (see this post and this one). In this experiment, you'll answer some questions about different types of emotions. I'll use this information to help design some upcoming verb-learning experiments.

As usual, the experiment is short and should take 3-5 minutes.


[snappy caption goes here]


-----
photo credit here

Learning the passive

If Microsoft Word had its way, passive verbs would be excised from the language. That would solve children some problems, because passive verbs are more difficult to learn than one might think, because not all verbs passivize. Consider:

*The bicycle was resembled by John.
*Three bicycles are had by John.
*Many people are escaped by the argument.


The bicycle was resembled by John: A how-to guide.

So children must learn which verbs have passives and which don't. I recently sat down to read Pinker, Lebeaux and Frost (1987), a landmark study of how children learn to passivize verbs. This is not a work undertaken lightly. At 73 pages, Pinker et al. (1987) is not Steve Pinker's longest paper -- that honor goes to his 120-page take-down of Connectionist theories of language, Pinker and Prince (1988) -- but it is long, even for psycholinguistics. It's worth the read, both for the data and because it lays out the core of what become Learnability and Cognition, one of the books that has had the most influence on my own work and thinking.

The Data


The authors were primarily interested in testing the following claim: that children are conservative learners and only passivize verbs that they have previously heard in the passive. This would prevent them from over-generating passives that don't exist in the adult language.

First, the authors looked at a database of transcriptions of child speech. A large percentage of the passive verbs they found were passives the children couldn't possibly have heard before because they aren't legal passives in the adult language:

It's broked? (i.e., is it broken?)
When I get hurts, I put dose one of does bandage on.
He all tieded up, Mommy.

Of course, when we say that the child couldn't have heard such passives before, you can't really be sure what the child heard. It just seems unlikely. To more carefully control what the child had heard, the authors taught children of various ages (the youngest group was 4 years old) made-up verbs. For instance, they might demonstrate a stuffed frog jumping on top of a stuffed elephant and say, "Look, the frog gorped the elephant." Then they would show the elephant jumping on top of a mouse and ask the child, "What happened to the mouse?"

If you think "gorp" has a passive form, the natural thing to do would be to say "The mouse was gorped by the elephant." But a child who only uses passive verbs she has heard before would refuse to utter such a sentence. However, across a range of different made-up verbs and across four different experiments, the authors found that children were willing -- at least some of the time -- to produce these new passive verbs. (In addition to production tests, there were also comprehension tests where the children had to interpret a passivization of an already-learned verb.)

Some Considerations

These data conclusively proved that children are not completely conservative, at least not by 4 years of age (there has been a lot of debate more recently about younger children). With what we know now, we know that the conservative child theory had to be wrong -- again, at least for 4 yos -- but it's worth remembering that at the time, this was a serious hypothesis.

There is a lot of other data in the paper. Children are more likely to produce new passive forms as they get older (higher rates for 5 year-olds than 4 year-olds). They taught children verbs where the agent is the object and the patient is the subject (that is, where The frog gorped the elephant means "the elephant jumped on top of the frog"). Children had more difficulty passivizing those verbs. However, a lot of these additional analyses are difficult to interpret because of the small sample sizes (16 children and only a handful of verbs per experiment or sub-experiment).

Theory 

Fair warning: the rest of this post is pretty technical.

What excites me about this paper is the theoretical work. For instance, the authors propose a theory of linking rules that have strong innate constraints and yet still some language-by-language variation.
The linkages between individual thematic roles in thematic cores and individual grammatical functions in predicate-argument structures is in turn mediated by a set of unmarked universal linking rules: agents are mapped onto subjects; patients are mapped onto objects; locations and paths are mapped onto oblique objects. Themes are mapped onto any unique grammatical function but can be expressed as oblique, object or subject; specifically, as the 'highest' function on that list that has not already been claimed by some other argument of the verb.
With respect to passivization, what is important is that only verbs which have agents as subjects are going to be easily passivized. The trick is that what counts as an 'agent' can vary from language to language.
It is common for languages to restrict passivized subjects to patients affect by an action ... The English verbal passive, of course, is far more permissive; most classes of transitive verbs, even those that do not involve physical actions, have the privilege of passivizability assigned to them. We suggest this latitude is possible because what counts as the patient of an action is not self-evident ... Languages have the option of defining classes in which thematic labels are assigned to arguments whose roles abstractly resemble those of physical thematic relations...
This last passage sets up the core of the theory to be developed in Learnability and Cognition. Children are born knowing that certain canonical verbs -- ones that very clearly have agents and patients, like break -- must passivize, and that a much larger group of verbs in theory might passivize, because they could be conceived of as metaphorically having agents and patients. What they have to learn is which verbs from that broader set actually do passivize. Importantly, verbs come in classes of verbs with similar meanings. If any verb from that set passivizes, they all will.

This last prediction is the one I am particularly interested in. A later paper (Gropen, Pinker, Hollander, Goldberg & Wilson, 1989) explored this hypothesis with regards to the dative alternation, but I don't know of much other work. In general, Learnability and Cognition go less attention than it should have, perhaps because by the time it was published, the Great Past Tense Debate had already begun. I've often thought of continuing this work, but teaching novel verbs to children in the course of an experiment is damn hard. Ben Ambridge has recently run a number of great studies on the acquisition of verb alternations (like the passive), so perhaps he will eventually tackle this hypothesis directly.

----
Pinker S, Lebeaux DS, and Frost LA (1987). Productivity and constraints in the acquisition of the passive. Cognition, 26 (3), 195-267 PMID: 3677572

Mendeley -- Not quite ready for prime time

Prompted by Prodigal Academic, I decided to give Mendeley a shot. That is, instead of working on a long over-due draft of a paper.

Mendeley is two things. First, it is a PDF library/reader. Second, it is a citation manager.

Currently, I used Papers for the first and Endnote for the second.  Both work well enough -- if not perfectly -- but it is a pain that I have to enter every paper I want to cite into two different programs.

(Don't tell me I could export my Papers citations library to Endnote. First, I'd have to do that every time I update my library, which is annoying. Second, Papers was created by someone who clearly never cites books, book chapters, conference proceedings, etc. So I'd have to fix all of those in Endnote ... every time I export.)

(Also, don't tell me about Zotero. Maybe it's gotten better in the last year since I tried it, but it was seriously feature-deficient and buggy beyond all belief.)

First glance

At first, I was pleasantly surprised. Unlike Papers, Mendeley is free so long as you don't want to use their Cloud functionality much (I don't). Papers is convinced there are people named Marc Hauser, Marc D Hauser, M D Hauser, and M Hauser. Mendeley can be led astray but has some nice options to allow you to collapse two different author records -- or two different keywords.

(On that note, my Papers library has implicit causality, Implicit causality and Implicit Causality all as different keywords. Once Papers has decided the keyword for a paper is, say, Implicit Causality, nothing on G-d's green Earth will convince it to switch to implicit causality. And its searches are case sensitive. Mendeley has none of these "features.")

Also, Mendeley will let you annotate PDFs and export the PDFs with your annotations in a format readable by other PDF viewers (if, for instance, you wanted to share your annotated PDF with someone). That's a nice feature.

These would all be nice additional features if the the core functionality of Mendeley was there. I'm sorry to say that the product just doesn't seem to be ready for prime time.
I typed "prime time" into Flickr, and this is what it gave me. Not sure why.
photo credit here.

Second glance

The first disappointment is that Mendeley does not have smart collections. Like smart playlists in iTunes, smart collections are collections of papers defined by various search terms. If you have a smart collection that indexes all articles with the keywords "implicit causality," "psych verbs" and "to read", then whenever you add a new paper with those keywords, they automatically go into the smart collection. This is very handy, and it's an excellent feature of Papers (except that, as mentioned above, my smart folder for implicit causality searches for the keywords "implicit causality," "Implicit causality" OR "Implicit Causality").

I suspect Mendeley doesn't have smart collections because it doesn't have a serious search function. You can search for papers written by a given author or with a given keyword, but if you want to search for papers written by the conjunction of two authors or any paper on "implicit causality" written by Roger Brown, you're out of luck. Rather, it'll perform the search. It just won't find the right papers.

Third glance

That might be forgivable if the citation function in Mendeley was usable. The idea is that as you write a manuscript, when you want to cite, say, my paper on over-regularization (18 citations and counting!), you would click on a little button that takes you to Mendeley. You find my paper in your PDF library, click another button, and (Hartshorne & Ullman, 2006) appears in your Word document (or NeoOffice or whatever) and the full bibliographic reference appears in your manuscript's bibliography. You can even choose what citation style you're using (e.g., APA).


Sort of. Let's say you want to cite two different papers by Roger Brown and Deborah Fish, both published in 1983 (which, in fact, I did want to do). Here's what it looks like:
Implicit causality effects are found in both English (BrownFish, 1983) and Mandarin (BrownFish, 1983)
At least in APA style, those two papers should be listed as (BrownFish, 1983a) and (BrownFish, 1983b), because obviously otherwise nobody has any idea which paper you are citing.

This gets worse. Suppose you wrote:
Implicit causality effects have been found in multiple languages (BrownFish, 1983; BrownFish, 1983).
Correct APA 5th Ed. style is, I believe, (BrownFish, 1983a, 1983b). Actually, I'm not sure what exactly the correct style is, because Endnote always takes care of it for me.

There are other issues. Mendeley doesn't have a mechanism for suppressing the author. So you end up with:
As reported by Brown and Fish (BrownFish, 1983; BrownFish, 1983), verbs have causality implicit in their meaning.
instead of
 As reported by Brown and Fish (1983a, 1983b), verbs have causality implicit in their meaning.
Nor does Mendeley know about et al:
Hauser, Chomsky and Fitch (Hauser, ChomskyFitch, 2001) put forward a new proposal....blah blah blah...as has been reported several times in the literature (Hauser, ChomskyFish, 2001; BrownFish, 1983; BrownFish, 1983).
That is, the second time you cite a paper with more than 2 authors, it doesn't contract to (Hauser et al. 2001). Unfortunately, there is no work-around for any of these problems. In theory, you can edit the citations to make them match APA style. Within a few seconds, a friendly dialog box pops up and asks you if you really want to keep your edited citation. You can click "OK" or click "cancel," but either way it just changes your carefully-edited citation back to its default -- at least it does on my Mac (the forums suggest that this works for some people).

It's possible that people who don't use APA won't have as many of these problems. Numbered citations, for instance, probably work fine. I've never submitted a paper anywhere that used numbered citations, though. So I either need to switch professions or continue using Endnote to write my papers.

Hopefully

One can hope that Mendeley will solve some of these issues. I found discussions on their "suggested features" forum going back many months for each of the problems discussed above, which suggests I may be waiting a while for these fixes. I do understand that Mendeley is technically in beta testing. But it's been in beta testing for over two years, so that's not really an excuse at this point.

Alternatively, maybe Papers will add a good citation feature (and discover books). Or maybe Zotero will confront its own demons. I'm going to have to wait and see.

It makes one appreciate Endnote. Yes, it's a dinosaur. No, it hasn't added any really useable features since I started using it in 2000. But it worked then, and it still works now. There's something to be said for that.

Findings: Which of my posts do you like best?

It will surprise nobody that I like data. By extension, it should surprise nobody that what I like about blogging is getting instant feedback on whether people found a post interesting and relevant or not. This is in contrast to writing a journal article, where you will wait minimally a year or two before anyone starts citing you (if they ever do).

How I feel about data.

Sometimes the results are surprising. I expected my posts on the suspicious data underlying recent graduate school rankings to make a splash, but the two posts together got a grand total of 2 comments and 16 tweets (some of which are automatically generated by FieldofScience). I didn't expect posts on my recent findings regarding pronoun processing to generate that much interest, but they got 6 comments and 26 tweets, putting them among the most popular, at least as far as Twitter is concerned.

To get a sense of which topics you, dear readers, find the most interesting, I compiled the statistics from all my posts from the fall semester and tabulated those data according to the posts' tags. Tags are imperfect, as they reflect only how I decided to categorize the post, but they're a good starting point.

Here are the results, sorted by average number of retweets:


label #Posts...Tweets_Ave... Reddit_Ave... Comments_Ave...
findings 2 13 0 3
publication 3 13 5 5
peer review 4 12 13 10
universal grammar 5 10 2 8
pronouns 3 10 0 2
GamesWithWords.org 2 9 0 1
scientific methods 7 8 7 7
neuroscience 1 8 0 5
overheard 1 7 0 1
language development 2 7 0 7
Web-based research 6 7 0 1
science and society 3 6 1 6
language 6 6 1 3
education 2 6 0 1
journalism 2 6 18 9
politics 7 6 0 2
science blogging 2 6 1 2
language acquisition 1 5 0 0
recession 2 5 1 3
the future 1 5 0 0
vision 1 5 0 1
graduate school 4 5 0 3
science in the media 3 5 12 7
method maven 2 5 18 10
media 3 4 0 1
psychology career path 1 4 0 2
lab notebook 3 3 0 1
none 4 3 0 0









Since we all know correlation = causation, if I want to make a really popular post, I should label it "findings, publication, peer review". If I want to ensure it is ignored, I shouldn't give it a label at all.

At this point, I'd like to turn it over the crowd. Are these the posts you want to see? If not, what do you want to read more about? Or if you think about your favorite blogs, what topics do you enjoy seeing on those blogs?

The Y-2011 bug

Because of accidentally typing "2010" on a few posts that were written in advance, today's post and a few others won't show up correctly in Reader or Twitter. Some people will have seen these posts, but some not.  Today's I think is particularly worth reading, so check it out here.

So maybe reading *should* be harder

Some weeks back I chided Jonah Lehrer for his assertion that he'd
love [e-readers] to include a feature that allows us to undo their ease, to make the act of reading just a little bit more difficult. Perhaps we need to alter the fonts, or reduce the contrast, or invert the monochrome color scheme. Our eyes will need to struggle, and we’ll certainly read slower, but that’s the point: Only then will we process the text a little less unconsciously, with less reliance on the ventral pathway. We won’t just scan the words – we will contemplate their meaning.
This sounded like a bunch of neuro-babble to me, partly because the research he cited seemed to be about something else entirely.

Obviously, the ventral pathway is the problem.

Spoke too soon


To the rescue come Diemand-Yauman, Oppenheimer & Vaughan, who just published a new paper in my favorite journal, Cognition. The abstract says it all:
Previous research has shown that disfluency -- the subjective experience of difficulty associated with cognitive operations -- leads to deeper processing. Two studies explore the extent to which this deeper processing engendered by disfluency interventions can lead to improved memory performance. Study 1 found that information in hard-to-read fonts was better remembered that easier to read information in a controlled laboratory setting. Study 2 extended this finding to high school classrooms. The results suggest that superficial changes to learning materials could yield significant improvements in educational outcomes.
The first experiment involved remembering 21 pieces of information over a 15-minute interval, which while promising, has it's limitations. Here are the authors:
There are a number of reasons why this result might not generalize to actual classroom environments. First, while the effects persisted for 15 min, the time between learning and testing is typically much longer in school settings. Moreover, there are a large number of other substantive differences between the lab and actual classrooms, including the nature of materials, the learning strategies adopted, and the presence of distractions in the environment... Another concern is that because disfluent reading is, by definition, perceived as more difficult, less motivated students may become frustrated. While paid laboratory participants are willing to persist in the face of challenging fonts for 90 s, the increase in perceived difficulty may provide motivational barriers for actual students.
Or it could just make the students bored.

In a second, truly heroic study, the researchers talked a bunch of teachers at a public high school into sending them all their classroom worksheets and powerpoint slides. The researchers recreated two versions of these materials: one in an easy-to-read font and one in a difficult-to-read font. Each of the teachers taught at least two sections of the same course, so they were able to use one set of materials with one group of students and the other set with the another group. The classes included English, Physics, Chemistry and History.

Once again, the researchers found better learning with the hard-to-read fonts.

Notes and Caveats


The researchers seem open to a number of possibilities as to why hard-to-read fonts would lead to better learning:
It is worth noting that it is not the difficulty, per se, that leads to improvements in learning but rather the fact that the intervention engages processes that support learning.
Moreover, unlike Lehrer, they don't recommend making everything harder to read, learn or do:
Not all difficulties are desirable, and presumably interventions that engage more elaborative processes without also increasing difficulty would be even more effective at improving educational outcomes.
There is one obvious concern one might have about their Experiment 2: the teachers were blind to hypothesis, but not to condition. The authors attempt to wave this away but asserting that the teachers would likely make the wrong hypothesis (that learning should be worse when the font is hard), and thus any "experimenter" bias would be in the wrong direction. However, we have no way of knowing whether the teachers attempted to compensate for the hard-to-read materials by explaining thing better. In fact, the authors had no way of testing whether the teachers behaved similarly in both conditions.

That's not at all saying I think it was a bad study or shouldn't have been published. I think it's a fantastic study. I don't know how they roped those teachers into the project, but this is the kind of go-get-it science people should be practicing. The study isn't perfect or conclusive, but no studies are. The goal is simply to have results that are clear enough that they generate more research and new hypotheses.


-------
Connor Diemand-Yauman, Daniel M. Oppenheimer, and Erikka B. Vaughan (2011). Fortune favors the bold (and the italicized): Effects of disfluency on educational outcomes Cognition, 118, 111-115 : doi:10.1016/j.cognition.2010.09.012

Qing Wen!

In the process of encouraging more Americans to study Spanish rather than Mandarin, Nicholas Kristoff notes that in Mandarin

there are thousands of characters to memorize as well as the landmines of any tonal language. 
How true! How true! Kristoff shortly proves the latter point in more ways than one:
The standard way to ask somebody a question in Chinese is “qing wen,” with the “wen” in a falling tone. That means roughly: May I ask something? But ask the same “qing wen” with the “wen” first falling and then rising, and it means roughly: May I have a kiss?

Just one possible reaction if you use the wrong tone.

Kristoff is right, so long as you don't mind sounding like a speech synthesizer. The classic description of third tone is a falling tone followed by a rising tone, but in practice it is relatively rare to pronounce the second half (the rising tone), particularly in fluent speech (in Taiwan, anyway; China has a lot of regional variation in Mandarin, so I don't know whether this holds everywhere). Figuring out when to pronounce the full tone and when not to is just one of many issues L2 Mandarin speakers run into.

Actually, third tone is worse than I just suggested. Qing wen is actually a good example, because the qing is also in third tone. When there are two third tones in a row -- as there are in the qing3 wen3 that  means "may I ask you a question?" (I'm writing in the tones with numbers here) -- the first one is pronounced as if it were second tone (start low and rise high). So even though qing technically doesn't change, its pronunciation depends on which wen you are using.

If you have three or more third tones in a row (e.g., ni3 you3 hao3 gou3 gou3 ma0?), deciding which syllables will be pronounced as if they were second tone is a complicated issue. I'd explain it to you, but I don't actually know myself. I've been told you actually have some flexibility in what you do, but I'm not sure that wasn't just another way of saying, "Sorry, I can't really explain it to you."

Zipcar

I've been an advocate for and member of Zipcar since my wife and I moved to Boston four and a half years ago. For that time period, I thought Zipcar was in every way superior to owning a car. Until last week, anyway. Now I'm reconsidering the car ownership issue.

Own or rent? That's the question.

It begins

I was already unhappy early in the week, having discovered Zipcar had overcharged us $375 over the past few months. At the beginning of August, I had added my wife as a driver on our account (we'd always had just one driver to save on the yearly membership fee) and also upgraded us to a fixed $75/month plan (which has some added benefits), having noticed that we'd spent more than $75 pretty much every month for the last year. I confirmed carefully with the representative that we would only be billed $75 combined, not $75 each.

It turns out, the hapless representative, rather than simply putting two of us on one account, made two accounts and put us each on both, and then charged us each every month. I didn't notice earlier as the charges appeared on different accounts, and I thought we only had one account, but the credit card charges were looking suspiciously large. It took a series of emails and a phone call to get that straightened out. They eventually agreed to refund us the bogus charges "as a one-time courtesy." That's a direct quote.

Thursday


We had an overnight trip to our favorite New England B&B for Christmas weekend (seriously, this place is fantastic and has one of the best restaurants I've been to anywhere in the world).

Our room was even more charming than it looks. Yes, that's a working fireplace.

As usual, I booked a Zipcar for the purpose. I believe it was a Nissan Sentra, helpfully parked in our apartment building's garage. On Thursday, I got an email from Zipcar saying that due to an unforeseen circumstance, they were bumping us to a Civic in the Government Center garage.

The exterior of the building is distinctive.

I've rented cars once or twice from that location. During the day, it's ok. At night, it's spooky as hell. I'd say it's deserted, but there are occasionally roving bans of teens doing who knows what.


I couldn't find a good picture of the interior.
This illustration is true to the spirit of the place.


So I emailed Zipcar, explaining that I didn't really like returning cars to that location at night, so was there maybe another car nearby I could use. Or could the Civic be relocated somewhere more pleasant for a couple days?

Friday


I don't know if anybody read that email, since I never got a reply. I did get another email on Friday, though, saying that due to another unforeseen circumstance, my reservation had been moved to a Smart Car parked in Somerville (a 15-20 minute drive from where we live).

Did I mention that the B&B was a 3 hour drive away? I didn't really want to drive a Smart Car on the highway for 3 hours, and I wasn't sure our stuff would fit (I had planned on bringing skiis). So I emailed, saying if I it was a choice between a Civic parked in the Dungeon of Dispair or a Smart Car, I'd take my chances with the dungeon.

This time, I got a quick email saying there were no other cars available. So I called and explained my situation. A very polite representative explained that there really were no other cars nearby, but would I take a Mazda 3 in Arlington (3 suburbs out from Boston, where I live)? They'd pick up the cab fare. Either that, or I could have $200 towards some other form of transportation. Or I could cancel my reservation. $200 wasn't going to cover a standard car rental (I checked), and the B&B reservation was nonrefundable (plus we'd been looking forward to it), so we went with the car in Arlington.

Saturday

We took a cab out to Arlington in the morning. It took maybe 20 minutes (we helpfully live next to I-93, making getting out of the city easy -- Thank you, Tip O'Neil and Ted Kennedy) and cost $33.75.

We'd only just gotten back onto I-93 in the direction of Vermont when there was a loud pop under the car and it sounded like something was dragging. On inspection (I took the next exit and pulled over), there was some piece of plastic hanging loose. The plastic itself didn't look problematic, but I wasn't sure what it had previously been holding in place. So I called Zipcar.

The representative agreed that the car was not safe to drive and asked us to return the car to its original location, and could we perhaps take a Prius from Wellesley College instead? If we needed a cab ride, they'd cover the fare. I pointed out that (a) we were already running pretty late, and (b) a cab fare to Wellesley was going to be pretty serious, esp. on top of our cab fare to Arlington, so could I just drive the car to Wellesley, drop it off there and take the Prius. She said that wasn't possible, since when the mechanics came out to service the Mazda, they wouldn't know where it was. I said if that was the problem, I was happy to tell the mechanics that the car was in Wellesley.

She put me on hold.

After a brief wait, she came back on the line to apologize, saying she hadn't gotten permission to drop the car off in Wellesley. Did I want to make the switch anyway? Or there was another car in Salem, MA, if we wanted.

Witch trials: Popular entertainment in Salem, MA

I've actually wanted to see Wellesley for a while (I like college towns), so we went with Wellesley. We called yet another cab (have I mentioned this was Christmas? Not a lot of cabs wanting to go to a deserted college town) and went to Wellesley. That fare was $66.05, including a Christmas-appropriate tip.


Finally, we got the Prius and set off. We had been traveling for 2 1/2 hours and were now farther from our destination than when we started.


View Larger Map

Our itinerarary: A: Home, B: Mazda in Arlington, C: Roughly where the car broke, D: Back in Arlington, E: Wellesley College. Vermont is in the North.

Sunday


The rest of Saturday was pretty good, and the B&B was everything we remembered (dinner, which was, as always, excellent, included what may be the perfect bread, from Orchard Hill Breadworks). Sunday morning we heard rumors that a serious blizzard was heading our way, though it wasn't expected to be bad until evening (before we had left, the weather report had put the chance of snow at only 30%). We got a slightly earlier start than we had planned, stopped at a few places on the way back.

Somewhere around 3:30 or 4:00, we entered Massachusetts and it began to snow. The state had put up blizzard warnings on the roads, requesting everyone to get off the road and go home. If we had been going straight home -- as we would have were we driving that much-mourned Sentra -- that wouldn't have been a problem. But we were going to Wellesley.

At least, we tried. As we neared Wellesley, the snow got very bad, and I frankly wasn't that comfortable driving, particularly once we left the highway and the streets weren't as well-plowed. My wife called a cab company to make sure we could get a ride back to Boston. They agreed to take us, but then called back shortly thereafter, saying that nobody was driving anywhere, didn't we know there was a blizzard going on? I think they made the right decision.

I did know there was a commuter rail station in Wellesley. We called the MBTA to see if the trains were still running. They said the trains would most likely run, but with significant delays. The next one wouldn't be for 3 hours. Oh, and the station we'd be waiting at is outside. In a blizzard.

As a backup plan, we checked to see if there was anywhere we could stay in the night in Wellesly. However, as Wellesley doesn't have any hotels, there appeared to be only one option:

The only available room in Wellesley last night.

We called up Zipcar to consult. They agreed to let us leave the car in Boston at no charge, as long as we told them where the car was. I'll give them credit for that decision, at least. We drove home, very slowly.

Monday


I wish I could say the story was over, but this morning I checked my email and saw that we were billed, not only for the Prius, but also for the Mazda 3 (the one that broke down). Plus, there was a late fee for returning the Mazda late. It seems that when the representative switched our rental from the Mazda to the Prius, she did so before we actually got back to Arlington. I sent in another email this morning. We can hope that they'll remove those charges "as a one-time courtesy."

I realize that owning a car has its own hassles. I don't expect Zipcar to be perfect, either. Everyone's allowed a bad week. As long as it's just one week. And as long as they give me my money back.

What did I learn from this experience? What I learned -- and what you should take home from this as well -- is to get your bread from Orchard Hill. Because it is fucking awesome bread.

*Update: Tuesday*


This morning I got a call from someone higher up in customer service at Zipcar, who listened to the whole story. She took the numbers for the cab rides in order to reimburse us directly, rather than my having to send in the receipts, which was nice. She also comped the entire weekend trip and added a $50 driving credit, which I also appreciate. The part I cared about more was that she at least seemed very interested in improving the service such that such problems would not be repeated or would be mitigated more quickly when they do. If this reflects a real commitment to efficient service, then hopefully this last week is an aberration, and we'll be able to go back to trusting and relying on Zipcar, as we have in the past.

*Another Update: Tuesday*


Now the vice president for member services has called to apologize in addition. It's great that they take this stuff seriously. I was going to take a temporary break from Zipcar and use a regular rental car company for some upcoming stuff, but now I think I'll give them another shot.

Crowdsourcing My Data Analysis

I just finished collecting data for a study. Do you want to help analyze it?

Puns

What makes a pun funny? If you said "nothing," then you should probably skip this post. But even admirers of puns recognize that while some are sublime, others are ... well, not.

Over the last year, I've been asking people to rate funniness of just over 2300 different puns. (Where did I get 2300 puns? The user-submitted site PunoftheDay. PunoftheDay also has funniness ratings, but I wanted a bit more control over how the puns were rated and who rating them.).

Why care what makes puns funny?

There are three reasons I ran this experiment. I do mostly basic research, and while I believe in its importance and think it's fun, the idea of doing a project I could actually explain to relatives was appealing. I was partly inspired by Zenzi Griffin's 2009 CUNY talk reporting a study she ran on why parents call their kids by the wrong names (typically, calling younger children by elder children's names), work which has now been published in a book chapter.

Plus, I was just interested. I mean: puns!

Finally, I was beginning a line of work on the interpretation of homophones. One of the best-established facts about homophones is that we very rapidly suppress context-irrelevant meanings of words -- in fact, so rapidly that we rarely even notice. If your friend said, "I'm out of money, so I'm going to stop by the bank," would you really even notice considering that bank might mean the side of a river?

A river bank. 
photo: Istvan, creative commons 

A successful pun, on the other hand, requires that at least two meanings be accessed and remain active. In some sense, a pun is homophone processing gone bad. By better understanding puns, I thought I might get some insight into language processing.

Puntastic


As already mentioned, my first step down this road was to collect funniness ratings for a whole bunch of puns. I popped them into a Flash survey, called it Puntastic, and put it on the Games With Words website. The idea was to mine the data and try to find patterns which could then be systematically manipulated in subsequent experiments.

It turns out that there are a lot of ways that 2300 puns can be measured and categorized. So while I have a few ideas I want to try out, no doubt many of the best ones have not occurred to me. Data collection was crowdsourced, and I see no reason why the analyses shouldn't be as well.

I have posted the data on my website. If you have some ideas about what might make one pun funnier than another -- or just want to play around with the data -- you are welcome to it. Please post your findings here.

If you are a researcher and might use the data in an official publication, please contact me directly before beginning analysis (gameswithwords$at*gmail.com) just so there aren't misunderstandings down the line. Failure to get permission to publish analyses of these data may be punished by extremely bad karma and/or nasty looks cast your way at conferences.

The results so far...

Unfortunately for the crowd, I've already done the easiest analyses. The following are based on nearly 800 participants over the age of 13 who listed English as both their native and primary languages (there weren't enough non-native English speakers to conduct meaningful analyses on their responses).

The average was 2.6 stars out of 7 (participants could choose anywhere from 1 to 7 stars, as well as "I don't get it," which was scored as -1 for these analyses), which says something either about the puns I used or the people who rated them.

First I looked at differences between participants to see if I could find types of people who like puns more than others. There was no significant difference in overall ratings by men or women.



I also asked participants if they thought they had good or poor social skills. There was no significant difference there, either.



I also asked them in they had difficulty reading or if they had ever been diagnosed with any psychiatric illnesses, but neither of those factors had any significant effect either (got tired of making graphs, so just trust me on this one).

The effect of age was unclear.


It was the case that the youngest participants produced lower ratings than the older participants (p=.0029), which was significant even after a conservative Bonferroni correction for 15 possible pairwise comparisons (alpha=.0033). However, the 10-19 year-olds' ratings were also significantly lower than the 20-29 year-olds' (p=.0014) and the 30-39 year-olds' (p=.0008), but obviously this was not true of the 40-49 year-olds' or 50-59 year-olds' ratings. So it's not clear what to make of that. Given that the overall effect size was small and that this is an exploratory analysis, I wouldn't make much of the effect without corroboration from an independent data set.

The funniest puns

The only factor I've looked at so far that might explain pun funniness is the length of the joke. I considered only the 2238 puns for which I had at least 5 ratings (which was most of them). I asked whether there might be a relationship between the length of the pun and how funny it was. I could imagine this going either way, with concise jokes being favored (short and sweet) or long jokes having a better lead-up (the shaggy dog effect). In fact, the correlation between pun ratings and length in terms of number of characters (r=.05) or in terms of number of words (r=.05) were both so small I didn't bother to do significance tests.

I broke up the puns into five groups according to length to see if maybe there was a bimodal effect (shortest and longest jokes are funniest) or a Goldilocks effect (average-length jokes are best). There wasn't.

In short, I can't tell you anything about what makes some people like puns more than others, or why people like some puns more than others. What I can tell you is which puns people did or didn't like. Here are the top 5 and bottom 5 puns:

1. He didn't tell his mother that he ate some glue. His lips were sealed.
2. Cartoonist found dead in home. Details are sketchy.
3. Biologists have recently produced immortal frogs by removing their vocal cords. They can't croak.
4. The frustrated cannibal threw up his hands.
5. Can Napoleon return to his place of birth? Of Corsican.
...
2234. The Egyptian cinema usherette sold religious icons in the daytime. Sometimes she got confused and called out, 'Get your choc isis here!'
2235. Polly the senator's parrot swallowed a watch.
2236. Two pilgrims were left behind after their diagnostic test came back positive.
2237. In a baseball season, a pitcher is worth a thousands blurs.
2238. He said, "Hones', that is the truth', but I knew elide.

Ten points to anyone who can even figure out what those five puns are about. Mostly participants rated this as "I don't get it."

----------------------
BTW Please don't take from this discussion that there hasn't been any serious studies of puns. There have been a number, going back at least as far as Sapir of the Sapir-Whorf hypothesis, who wrote a paper on "Two Navaho Puns." There is a well-known linguistics paper by Zwicky & Zwicky and at least one computer model that generates its own puns. However, I know a lot less about this literature than I would like to, so if there are any experts in the audience, please feel free to send me links.

Paper submitted

I just submitted a new paper on pronoun resolution ("Do inferred causes really drive pronoun resolution"), in which I argue that a widely-studied phenomenon called "implicit causality" has been misanalyzed and is in fact at least two different phenomena (as described in this previous post). You can find the paper on my publications page. Comments are welcome.

I always find writing up methods and results relatively easy. The trick is fitting the research into the literature in a way that will make sense and be useful to readers. That is, while the narrow implications are often clear, it's not always obvious which broader implications are most relevant. That is, the paper has clear implications for the few dozen people who study implicit causality, but one would like people beyond that small group to also find the results relevant.

I tried a few different approaches before ultimately settling on a particular introduction and conclusion. I was curious how much the paper had changed from the first draft to the last.

Here's the first draft, according to Wordle:


Here's draft 2:



The most obvious differences is that I hyphenated a lot more in the final draft (I was trying to make the word limit). But it doesn't appear that the changes in theme -- as measured by Wordle -- were all that drastic. That's either a good sign (my paper didn't lose its soul in the process of editing) or a bad sign (I didn't edit it enough).

I guess we'll see when the reviews come back in.