May 16, 2015

So, What is Psy Like as a Keynote Speaker?

Posted in Uncategorized tagged at 7:56 pm by mariawolters

At the recent CHI 2015 conference on Human-Computer Interaction in the heart of Gangnam District, Seoul, South Korea, we had a somewhat unexpected keynote speaker: Park Jae-Sang, also known as Psy, probably most famous for a certain song called Gangnam Style.

Why did I say Park Jae-Sang spoke? Because that’s who came. No glasses, no kooky clothes, no props, no silly dance. Just a man in a business suit and a microphone.

Park Jae-Sang is an ambitious man who is very conscious of the image he projects. At the same time, he tries to be himself. His constant jokes about his weight made it clear that he chose to stay his chubby self, despite K-Pop pressures to be slender and beautiful. His constant jokes about his English (which was excellent) showed his insistence on his Korean roots.

The tale he told was not of a Social Media Ninja, but of a shrewd business man whose associates told him about this YouTube thing – they convinced him, so he decided to explore it.

Then, he explored fame, and what it means to have a global hit. He concluded that Gangnam Style was a one-off, and the way forward for him is to be who he is – Park Jae-Sang, composer, businessman, artist, Korean, uniquely himself.



April 29, 2014

Medication Reminders – What is the Minimal Effective Dose of Technology?

Posted in Uncategorized tagged , , , at 2:14 pm by mariawolters

As I type this, I am sitting on a park bench in Toronto’s Roundhouse park, a sanctuary for old steam and diesel engines in front of the state-of-the-art Metro Toronto Convention Centre.
I’m here with my daughter, who is currently busy repurposing a steam train playground installation as a boat. Behind me, traffic crawls along the Gardiner Expressway which cuts off the Waterfront and its condos and building sites from Downtown.

I am at the 2014 CHI conference on Computer-Human Interaction, which is one of the largest conferences on making technology useable.

On Monday morning, the conference opened with a very thought provoking keynote by Margaret Atwood on robots, technology, and humans. One of the many points she made was about unexpected perspectives on familiar technologies. Just as my daughter converted a train into a boat, technology is invented for one purpose, but then can serve many others. The true potential of a thing is an unknown unknown, in Donald Rumsfeld’s words. It’s a wide open space, limited only by creativity and serendipity.

At alt.chi (on Tuesday before lunch), I’m going to argue that what helps us remember to take our medications is not shiny new purpose built apps – rather, we need to delve into the unknowns and be creative, so that remembering medications is as little work as possible.

(The mathematically inclined readers among you can now imagine using your favourite approach to minimising a differential equation.)

As Juliet Corbin and Anselm Strauss argued in a series of seminal papers, there are three layers of work associated with illness. First, there’s the illness work proper – taking medications, doing prescribed exercises. Then, there’s everyday work – roughly, getting on with your life while being ill. Finally, there’s biographical work – work on your own identity and values. Not to mention that being ill means that you are drained and, by definition, not able to function at your best.

Illnesses create additional everyday and biographical work. Take people with diabetes. They need to schedule regular checkups with their health care providers, take prescribed pills, and remember to refill their prescriptions on time. They may need to overhaul the way they eat. This can mean spending more time preparing and sourcing foods that won’t aggravate their illness – so more everyday work. Finally, they need to come to terms with their diagnosis. Often, they will need medication for the rest of their lives. They need to cut back radically on cakes and sweets. What’s worse, in public discourse, people with diabetes are often stigmatised as fat slobs who ate themselves sick.

So, assume they forget their pills. Let’s just install a smartphone app, shall we? But what about people who struggle to work their phones (and they’re not all elderly)? What about people who rarely use their phones (again, they’re not all elderly)? There are many reasons why smartphone apps can and will fail – and a common denominator of many of them is that using those apps (indeed, using a smartphone) is too much work.

Work on top of work, while the person who has to do the work is not at their best.

Let’s be honest – how do you remember to take your medication? Do you use technology? If it works for you, great.

But what if it doesn’t?

What if the most effective dose of technology is not one app, but none?

People, we need to talk.


1. Corbin, J., and Strauss, A. Managing chronic illness at home: Three lines of work. Qualitative Sociology 8, 3 (1985), 224–247

Maria Wolters (2014). The Minimally Effective Dose of Reminder Tehcnology Proceedings of CHI 2014 – alt.chi DOI: 10.1145/2559206.2578878

June 27, 2012

Building a Network Means Knowing Who You Talk To

Posted in Uncategorized at 9:33 am by mariawolters

Why do people follow others on Twitter, only to require them to validate with a service like TrueTwit when they wish to follow back?

When I follow somebody, I make that decision based on interacting with them, having a look at their timeline, or finding them in follower lists of people I rate. I will also follow back everybody who follows me who is not a bot and who is not purely a professional social media marketer. (I don’t even care whether I speak their language. I’m a linguist, I love seeing Babel in my timeline.) I’m not online to promote my research or my business, I’m online to build relationships with people. My timeline is a wonderfully diverse crowd, from Tories to Greens, from strict Catholics to adamant atheists.

So if you take a considered decision to follow somebody, then why would you require the new addition to your timeline to confirm that they are not a bot? The only situation I can see is people adding followers automatically through online services. But if you add people automatically, that’s usually a sign that you don’t interact. And broadcasting is not what social media is about. It’s an interaction, as Robert Fondalo, perhaps the only marketer I have ever followed back, expresses this succinctly in this post

Looking at somebody’s timeline is also important because this, to me, is the primary context in which tweets need to be interpreted. Tweets are 140 character messages, so much of the meaning and nuance needs to be implied. It’s usually quite clear how to read a tweet after having seen the person tweet for a couple of weeks. Failing to take such context into account is one of the major sources of fights and hissy fits on Twitter. At its worst, it can even lead to a two and a half year fight with the courts when a tweet that went out to 600+ followers, a tweet whose author regularly jokes and banters with the people on his timeline, and a tweet for which the contextually appropriate reading is “(bad) joke”, is taken completely out of context.

Social media is not broadcasting. It’s talking to people. Know your audience, and get to know people a little before you follow. It’s that simple.

May 26, 2012

Science Defies Stereotypes

Posted in Uncategorized at 9:28 am by mariawolters

In a recent post, Athene Donald argued that an overemphasis on “geekiness” and associated cultural stereotypes might turn many girls off science.

I agree with her. Although I’m a proud geekette, there are many paths to science, and many ways of living as a scientist. You don’t have to conform to anybody’s stereotype; all you need is a passion for finding out how stuff works.
But I would widen that approach to all genders and youth subcultures.

Turned off because boys who are into science are weak dweebs? Look at this geologist, and this computer scientist (double amputee and expert climber).

Think that female scientists are colourless and boring? Look at this computer scientist.

Do you feel disdain for people with a scientific bent because they don’t get the humanities? Look at this feminist and exercise scholar and think again.

Whoever you are, whatever you are, there will be a scientist out there who is just like you.

May 22, 2012

How Can We Help Researchers For Whom English is a Second Language Write Good Papers?

Posted in Uncategorized at 10:18 am by mariawolters

Neuroskeptic pointed out one instance where plagiarism may be forgivable – when authors whose first language is not English copy small passages from papers to put together their literature review

Now, if authors are clever enough to write an academic paper, their English should be good enough to summarise complex papers in a few words, right?

Wrong. Writing in a foreign language is very difficult, and writing a complex text such as an academic paper in a foreign language is a highly specialised skill. Native speakers of English, most of whom would not be able to write a paper in another language, don’t realise just how difficult this is. My own English is passable, but then I won a Second Prize in the Federal German Foreign Language Contest (there were several first and several second prizes) with English as my primary competition language, have been living in Scotland for 10+ years. And I still make mistakes.

Writing skills need to be maintained I entered the Contest as a 17-year-old with French as my second language. For my Abitur (A-levels), I wrote a long essay about the writer and philosopher Albert Camus in French – today, I can hardly string a blog post together, even though I can still read and understand Camus just fine. In fact, English is the only foreign language in which I can write papers; I would be utterly out of my league if I had to write in Spanish or French.

Now imagine that you don’t spend a lot of time writing English. All of a sudden, you need to put together a paper in the language. Writing well in one’s mother tongue is hard; finding the right turns of phrase in a foreign language is even harder, especially when there are strict page limits, and your field does not have very rigid structures for academic papers.r What do you do?

And, most importantly, as a reviewer, how do you help authors who struggle with their writing?

My own strategy represents a trade-off between time required to review the papers I’ve taken on and diligence. I point out major errors, in particular where terms are used in ways that prevent an English speaker from understanding what is meant, but I let most of the small things slide, in particular when my verdict is “revise and resubmit”. I then provide detailed feedback on the resubmission.

Ideally, journals would have mentors or specialised, paid editors that can help people who struggle with writing English; in the absence of such resources, I often recommend that authors have their papers proofread by a native speaker of English. I know that this can come across as condescending, especially if the authors have worked very hard to write an acceptable paper.

So, what can we do to address this problem as a community? Turning a blind eye to small instances of plagiarism? There are a couple of other options that are relatively inexpensive

  • Develop clear language standards, and enforce them when reviewers whose native language is English expect literary masterpieces.
  • Put together links on field-specific English for Academic Purposes that authors can access.
  • Provide guidance on rewording results and findings from papers for the literature review that helps authors negotiate the line between reporting and plagiarism
  • Provide reviewers with the option of submitting annotated PDFs of the paper together with their review – it’s very cumbersome to make a long list of page, line, and paragraph numbers, copy the bad wording, and type out the correction, especially if line numbers do not line up properly with the lines in the text (or when there are no line numbers at all)

What are your suggestions?

May 15, 2012

Knowledge Pyramids in Telecare

Posted in Uncategorized at 12:26 pm by mariawolters

In circumstances where it can take clinicians and therapists a long time to reach the patients who need them, or where patients need to travel long distances in order to see a specialist, telemedicine comes into its own. Telecardiology or teleradiology allow specialists to receive and assess data from remote locations, giving instant feedback if the communication infrastructure is in place. Thus, telemedicine bridges a gap in knowledge and expertise.

This can be extended to medical education, as Mark Barr from Intel showed at the recent Med-e-Tel conference in Luxemburg. In his presentation, he drew up a knowledge pyramid, where specialists have the highest level of knowledge, followed by generalists, medical nurses, and health workers. Medical education was one of the ways of bridging this knowledge gap.

However, I think that there are really two knowledge pyramids – one of medical knowledge required to help the patient, and one of implementation knowledge required to make sure the patient can get the help they need and implement the required measures. Fitting them both together leads to a continuum where one source of knowledge increases and the other decreases.

For example, if the specialist recommends regular exercise, such as brisk walking for thirty minutes a day, the local health worker can tell people about good routes and point them to local walking groups, if they exist. If the specialist recommends a healthier diet, the health worker can help with suggestions of cheap, nutritious meals, local sources of good ingredients, or cookery classes.

The specialists do not need the local knowledge just like the health workers don’t need the specialist knowledge, but both ends of the continuum need to work together for best results. A top-down conceptualisation of telemedicine, where education just flows along the medical knowledge path, but not back along the implementation one, is – to my mind at least – deeply flawed.

May 8, 2012

Moodscope: Track your mood – social, but discreet

Posted in Uncategorized at 2:22 am by mariawolters

There are many low and high tech solutions for tracking one’s moods and feelings, from the humble notebook to the shiny app, from detailed, free-form diaries to ticking a couple of boxes on a form. Many people track their mood informally using social media, letting their online support network know how they feel through blog posts, tweets, and status updates.

Moodscope is a web application that allows users to log and share their mood. Users can rate how you are feeling using 20 mood and emotion adjectives, such as scared, guilty, ashamed, proud, or alert. For each adjective, users need to specify whether they feel like this very slightly or not at all (0), a little (1), quite a bit (2), and extremely (3). This can be done as often as the user wishes.

The interface design is &ldots; unique. For each adjective, the user sees a card. On one side, there are the numbers 3 and 2, on the other side 0 and 1. The user then turns and spins the card until the number that corresponds to the current intensity of that particular emotion is on the top. To log that number, users click on the verbal description. At the end, scores are converted into percentages, where a high percentage reflects good mood, and a low percentage bad mood. After the first three recorded values, a summary is added that describes trends verbally and comments on how frequently Moodscope is used. Daily emails with short, motivating texts serve as reminders. These emails are sent whether or not the user has already completed a mood log that day.

The social aspect is very discreet. Users can nominate buddies who will be sent their Moodscope scores. Buddies need to confirm that they are willing to receive the email updates. All the buddy sees is the current percentage, but there’s an opportunity to discuss scores privately on the Moodscope web site.

One User’s Experience

I started using Moodscope a few weeks ago; I have only used the free version, and my comments may not apply to the paid version.

What first struck me was the pared down functionality. You confirm today’s date, and then it’s straight onto rating the twenty adjectives. Operating the rating interface is slow and cumbersome. You need good eye sight to make out the letters against the background (cards are either reddish or blueish), and the writing can be difficult to read. You also need to plan the flips and turns required in order to get the number you need to enter. This is a strain, and it definitely deters me from logging my mood more than once a day. I am not sure how well a person would do who is easily discouraged, or whose mood makes focusing very difficult.

The summary is not very helpful. As I am logging, my mood shows large swings, but that mostly depends on variation in my mood throughout the day. Moodscope as it stands misses this zig-zag, because the summary only compares my scores to the all time maximum, all time minimum, and average. If there is a large dip, the summary suggests that specific events may have happened to change my mood, when this is nothing but my normal fluctuation.

Fortunately, Moodscope also has a great graphic display.
The graphical display of the change in mood over time is very useful – it allows you to zoom in and out, clearly shows the range of mood that has been registered over time, and can be adjusted to show shorter or longer time periods.

An important limitation is that Moodscope – for good reason – does not ask people whether they are thinking about suicide or self-harm. If it did, there would be unpleasant implications. First of all, it places unreasonable strain on the buddies who receive the reports. What if people indicate that they are ready to kill themselves, but this is not reported to the buddies? And if it is reported to the buddies, what if they discover the warning after the fact? However, this sensible limitation means that Moodscope can miss a significant improvement in cases where people feel still low, but no longer suicidal.

The main reason I would keep using Moodscope is the buddy function. It means that I can let others know how I am feeling through the relatively private medium of email. Otherwise, the interface would be too much of an annoyance, as registering one answer can mean up to three clicks, which needs to be done twenty times. There are thousands of applications diligently monitoring all public social media activity. Somehow, mood is a bit more private than that – it’s good that there’s an application which keeps it private.

May 3, 2012

Open Access for the Arts and Humanities

Posted in Uncategorized tagged , at 1:44 pm by mariawolters

On Twitter, Rebekah Higgit reacted to David Willetts’ speech on Open Access to research results by asking – well, what about arts and humanities?

In this short post, I would like to outline a couple of practices that I am familiar with from linguistics and psychology that do not require publisher action, but that can be implemented today, by any researcher, unless they’ve signed a particularly stifling publication contract.

  • Directly linking to copies of their work on their own web page. This is the most basic version.
  • Circulating drafts: Plenty of papers do the rounds in draft form before they are even submitted for peer review. This happened to a paper I worked on which ended up being published in the prestigious journal Language. If you look at the Google Scholar citation page, quite a few citations predate the actual publication. These come from circulated previous drafts.
  • E-mail request systems: Many groups working in psychology have an ingenious mechanism for getting around publisher’s restrictions. They set up a central email address whereby you can request a paper, and either a bot or a person monitoring that address will send you a version, either a pre-print as submitted to the publisher for typesetting, or the Authors’ Copy of the typeset version.
  • Pure online journals: I’m aware of a few well-regarded online journals in the humanities and social sciences, such as The Qualitative Report and Forum: Qualitative Research.

So, there are a number of ways of enabling Open Access now – let’s hope Willetts’ team manages to negotiate a model that allows more genuine open access for all branches of learning.

May 2, 2012

How a Post on Biased Thinking Used Research Twisted to Fit Its Own Bias

Posted in Uncategorized at 2:08 pm by mariawolters

Last week, New Statesman journalist Martha Gill published a column in the New Statesman’s Current Account blog about the way in which our own biases affect how we process information. She based her argument not just on her own observations, but on a scientific study.
Unfortunately the actual results of the study are not nearly as neat as Gill would like them to be.

Before we start, let me be clear: Studies can be distorted at any step from the reseachers’ fertile minds to the page or blog post, with the most common culprit being (misreported) press releases. I don’t know where the paper Gill cites got twisted beyond its results; I merely based my analysis on the way she reported the study.

Gill versus Nyhan/Reifler

A study published in the journal Political Behaviour shows just how reluctant people are to engage with facts that don’t support their world-view.

Notice that the reference is not given; for fact fans, it’s Nyhan, Brendan and Reifler, Jason (2010): When Corrections Fail: The Persistence of Political Misconceptions. Polit Behav (2010) 32:303–330. The paper is freely available online, as far as I can see. Ben Goldacre explains much better than me why this matters: It allows readers to check whether conclusions were reported correctly.

In the experiment, conducted in 2005, participants were given fake news stories.

There were two experiments, one in 2005 and one in 2006, that consisted of a total of four studies. This is very important – we will see later why.

These news stories were embedded with false facts: that tax cuts under the Bush administration increased government revenues, that weapons of mass destruction had been found in Iraq and that Bush had banned stem-cell research (he only limited some government funding).

The first experiment in autumn 2005 looked at correcting people’s impression that Iraq had weapons of mass destruction; the second in spring 2006 repeated the earlier study, and repeated it with two more topics, tax cuts and stem cell research. In the first experiment, the story with the “false facts” came from Associated Press, in the second experiment, half the participants were shown stories that purported to be from the New York Times, a notoriously liberal paper, and half saw exactly the same story, but this time, it supposedly came from Fox News, a notoriously conservative news source.

After each statement, the researchers put in an unambiguous correction – and then tested the participants to see if they picked this up.

The supposedly unambiguous correction was in fact a paragraph in the same story that reported findings from a relatively objective source which contradicted the key statement in the first paragraphs. So what we have here is not statement / counterstatement, but rather a classic “he said/she said” structure, where journalists present both views.

Here’s the original text from Experiment 1, together with the correction.

Wilkes-Barre, PA, October 7, 2004 (AP)—President Bush delivered a hard-hitting speech here today that made his strategy for the remainder of the campaign crystal clear: a rousing, no-retreat defense of the Iraq war. Bush maintained Wednesday that the war in Iraq was the right thing to do and that Iraq stood out as a place where terrorists might get weapons of mass destruction. ‘‘There was a risk, a real risk, that Saddam Hussein would pass weapons or materials or information to terrorist networks, and in the world after September the 11th, that was a risk we could not afford to take,’’ Bush said.

While Bush was making campaign stops in Pennsylvania, the Central Intelligence Agency released a report that concludes that Saddam Hussein did not possess stockpiles of illicit weapons at the time of the U.S. invasion in March 2003, nor was any program to produce them under way at the time. The report, authored by Charles Duelfer, who advises the director of central intelligence on Iraqi weapons, says Saddam made a decision sometime in the 1990s to destroy known stockpiles of chemical weapons. Duelfer also said that inspectors destroyed the nuclear program sometime after 1991.

[All subjects]
The President travels to Ohio tomorrow for more campaign stops.

(Nyhan & Reifler, 2010, p. 324f.)

Can you see how easy it is to reframe this as a piece that just presents two different points of view? Your evaluation of the correction will depend largely on your view of the CIA, Big Government, and bureaucrats who write reports.

They didn’t. Participants who identified themselves as liberal ignored the correction on stem-cell regulations and continued to believe Bush had issued a total ban. Conservatives not only ignored the corrections on Iraq and the tax cuts but clung even more tenaciously to the false information. Facts had made things even worse.

Well, what actually happened?

First of all, this was never about people actually changing their opinion; the researchers are clear that this is future work. Instead, this was a between-subjects design. Half the participants read the story with the correction before they answered the question, half read the story without the correction.

The question participants answered was designed to measure shift in opinion:

Question to participants:
Immediately before the U.S. invasion, Iraq had an active weapons of mass
destruction program, the ability to produce these weapons, and large stockpiles of WMD, but Saddam Hussein was able to hide or destroy these weapons right before U.S. forces arrived.

  • Strongly disagree [1]
  • Somewhat disagree [2]
  • Neither agree nor disagree [3]
  • Somewhat agree [4]
  • Strongly agree [5]

    (Nyhan & Reifler, 2010, p. 325)

    The authors looked at four predictors of opinion: whether participants had seen the correction, what their ideology was (on a scale from liberal to conservative), how much they knew about politics, and whether their ideology affected their reaction to the correction. When they used all of these predictors to model participants’ answer to the question, by far the largest effect was political knowledge. In the first study, they also found a clear effect of ideology. The correction was more likely to work the more liberal participants were; it backfired for conservatives. Conservatives who read the corrected text were more likely to believe that Iraq indeed had Weapons of Mass Destruction than Conservatives who didn’t.

    In the second experiment in Spring 2006, where they looked at the Weapons of Mass Destruction issue again, that backfire effect for conservatives was gone. The only way they could replicate it was to look at a small subset of their participants who had stated that Iraq was most important for them, but that’s a post-hoc analysis (in other words, fishing for results). There are many possible explanations for this, but the authors argue that the main reason was the change in public opinion between Autumn 2005 and Spring 2006. So, far from showing that journalists are powerless, the results actually suggest that persistent corrections across a range of media might be far more powerful than a single story in a point/counterpoint scenario.

    When the researchers used a text about tax cuts, however, they saw the same “backfire effect” in conservatives that they had observed in the Autumn 2005 study. In effect, strong conservatives who read the text with the correction were more likely to believe the incorrect fact than strong conservatives who didn’t see the correction. When Nyhan and Reifer looked at liberals’ reactions to a text about stem cell effects, there was no such backfire effect; liberals who read the correction were less likely to accept the correction, but they didn’t show an increased conviction that the opposite was true.

    What about people who are neither strongly liberal or strongly conservative? Well, in all of the four studies Nyhan and Reifer conducted, the effect of ideology was gradual, i.e. Centrists were far less likely to be influenced by ideology than either people with strong liberal or strong conservative views.

    (By the way, all participants in this study were undergraduates at a Catholic university; the researchers say that their study needs to be repeated with a more representative sample of the population.)

    So, What Are We To Make Of This?

    People’s perceptions can change, but they don’t change based on reading a single contradictory story, even if it comes from supposedly trustworthy sources. (In fact, in the second experiment, the purported source of the story (New York Times vs Fox News) did not make any difference to the results at all).

    What we need to do is understand why people persist in their beliefs despite contrary evidence. In particular, we need to look at

    • how much people already know (which explained a lot of the variation in the data)
    • how persistent their beliefs are
    • what motivates their reasoning

    So what can journalists do? Quite a bit, as it turns out.

    First of all, once a belief has been formed, it tends to persevere. So, those journalists and news media who break stories have a unique opportunity to affect people’s beliefs about the issues they report, because first impressions are likely to stick (c.f. Ross and Lepper’s (1980) work on belief perseverance, cited after Nyhan & Reifler 2010).
    Secondly, and most importantly, if misinformations are persisting, keep correcting them, and seek as many allies as possible to spread the correct information. People who are ‘‘confronted with information of sufficient quantity or clarity… should eventually acquiesce to a preference-inconsistent conclusion.’’ (Ditto & Lopez, 1992, p. 570).

    So if you are a journalist or an activist, and Gill’s column discouraged you, take heart. Change happens, but it takes patience and persistence. Keep going!

    Additional References

    Ditto, P. H., & Lopez, D. F. (1992). Motivated skepticism: use of differential decision criteria for preferred and nonpreferred conclusions. Journal of Personality and Social Psychology, 63(4), 568–584.

    Ross, L., & Lepper, M. R. (1980). The perserverance of beliefs: Empirical and normative considerations. In R. A. Shweder (Ed.), Fallible judgment in behavioral research: New directions for methodology of social and behavioral science (Vol. 4, pp. 17–36). San Francisco: Jossey-Bass.

April 19, 2012

The Academic Spring Debate – Towards Open Access for All

Posted in Uncategorized at 8:41 pm by mariawolters

Both Jo Brodie and Alice Bell have questioned the wider implications of the “Academic Spring” of open access research papers. Providing easy access to research papers is all well and good, but what about helping people make sense of them?

Jo raises an issue that is particularly important, patients and informal carers researching their own condition on the internet. How are they going to cope with medical journal articles, which are written to tight word counts using rigid structures and highly technical vocabulary? Even doctors need training in those skills!

Alice and Jo suggest that journal papers themselves should be written in a more accessible style. While that would be ideal, there is a fundamental problem with this from a communication theoretic perspective. Text structures, conventions of language use, and technical vocabulary emerge because they allow people who share similar background knowledge to communicate efficiently. Sometimes, the most efficient code is graphical; sometimes, it is mathematical.

To be clear, I am not defending badly written papers. Good, clear, engaging writing is still very important, as anybody who has reviewed a badly written paper at 10pm in a desperate bid to make a review deadline will be able to attest. Good writers use technical terms to make their point succinctly, to indicate their theoretical or methodological background, and to link into a wider body of work on that area.

While it is often possible to explain relevant concepts to lay people, this requires changes to the structure of the text, to ensure concepts are properly defined and anchored before they are used. Another problem is making a judgement call about the level of detail that needs to be explained. Let’s take significance testing. How do we communicate what a p-value of 0.05 actually means in practice? And are p-values the right thing to report? What other background should be given?

Fortunately, there is a way out of this conundrum which should be familiar to most researchers who have written or reported on a research grant – the lay summary. Lay summaries are not an easy way out. In order to write a good summary, you need to boil down your research to its essential goal and its main results. This requires a deep understanding of one’s own results. What have we done? What does it actually mean? What more do we need to know?

Traditionally, abstracts and keywords were deemed to be enough to help readers situate the research reported in a paper. Many journals go even further than this. Elsevier journals encourage graphical abstracts and one-sentence research highlights. Many medical journals require a box that summarises the main contribution of the paper to the state-of-the-art; some even use prompts to help authors structure such a text. Some journals have (or had) English abstracts translated into Spanish, French, or German, to help researchers for whom English is not their native language.

From these traditions and innovations, it is but a small step to short summaries written specifically for a non-specialist audience. Funders have shown the way in developing a standard structure, issuing writing guidelines, and providing examples. Unfortunately, what is still missing is a culture of good editing, which costs time and money, but is crucial, especially for lay summaries. The SPARC initiative led the way here, with professionally crafted texts (here’s an example of a project on auditory reminders that I led), but so far, none of the funders I have written summaries for have followed that shining example.

But who then helps non-specialist readers to critically assess what they are reading? Here is where science blogging comes into its own and where science communicators like Ben Goldacre, Petra Boynton, Ed Yong and Kate Clancy shine, to name but a tiny handful. Bewildered by the sheer variety of science blogs? Try Science Seeker or Research Blogging, from there, go to the big and small networks. By now, there is a critical mass of science bloggers who have been honing their craft for years. Add to that well-crafted lay summaries, and we might be onto a science communication winner.

Edited to add (April 20): Stuart Cantrill alerted me to the fact that graphical abstracts have a long history in chemistry. Here is a link to an editorial on graphical abstracts that he has written for Nature Chemistry (requires registration).

Next page