June 27, 2012
Why do people follow others on Twitter, only to require them to validate with a service like TrueTwit when they wish to follow back?
When I follow somebody, I make that decision based on interacting with them, having a look at their timeline, or finding them in follower lists of people I rate. I will also follow back everybody who follows me who is not a bot and who is not purely a professional social media marketer. (I don’t even care whether I speak their language. I’m a linguist, I love seeing Babel in my timeline.) I’m not online to promote my research or my business, I’m online to build relationships with people. My timeline is a wonderfully diverse crowd, from Tories to Greens, from strict Catholics to adamant atheists.
So if you take a considered decision to follow somebody, then why would you require the new addition to your timeline to confirm that they are not a bot? The only situation I can see is people adding followers automatically through online services. But if you add people automatically, that’s usually a sign that you don’t interact. And broadcasting is not what social media is about. It’s an interaction, as Robert Fondalo, perhaps the only marketer I have ever followed back, expresses this succinctly in this post
Looking at somebody’s timeline is also important because this, to me, is the primary context in which tweets need to be interpreted. Tweets are 140 character messages, so much of the meaning and nuance needs to be implied. It’s usually quite clear how to read a tweet after having seen the person tweet for a couple of weeks. Failing to take such context into account is one of the major sources of fights and hissy fits on Twitter. At its worst, it can even lead to a two and a half year fight with the courts when a tweet that went out to 600+ followers, a tweet whose author regularly jokes and banters with the people on his timeline, and a tweet for which the contextually appropriate reading is “(bad) joke”, is taken completely out of context.
Social media is not broadcasting. It’s talking to people. Know your audience, and get to know people a little before you follow. It’s that simple.
May 26, 2012
I agree with her. Although I’m a proud geekette, there are many paths to science, and many ways of living as a scientist. You don’t have to conform to anybody’s stereotype; all you need is a passion for finding out how stuff works.
But I would widen that approach to all genders and youth subcultures.
Turned off because boys who are into science are weak dweebs? Look at this geologist, and this computer scientist (double amputee and expert climber).
Think that female scientists are colourless and boring? Look at this computer scientist.
Do you feel disdain for people with a scientific bent because they don’t get the humanities? Look at this feminist and exercise scholar and think again.
Whoever you are, whatever you are, there will be a scientist out there who is just like you.
May 22, 2012
Neuroskeptic pointed out one instance where plagiarism may be forgivable – when authors whose first language is not English copy small passages from papers to put together their literature review
Now, if authors are clever enough to write an academic paper, their English should be good enough to summarise complex papers in a few words, right?
Wrong. Writing in a foreign language is very difficult, and writing a complex text such as an academic paper in a foreign language is a highly specialised skill. Native speakers of English, most of whom would not be able to write a paper in another language, don’t realise just how difficult this is. My own English is passable, but then I won a Second Prize in the Federal German Foreign Language Contest (there were several first and several second prizes) with English as my primary competition language, have been living in Scotland for 10+ years. And I still make mistakes.
Writing skills need to be maintained I entered the Contest as a 17-year-old with French as my second language. For my Abitur (A-levels), I wrote a long essay about the writer and philosopher Albert Camus in French – today, I can hardly string a blog post together, even though I can still read and understand Camus just fine. In fact, English is the only foreign language in which I can write papers; I would be utterly out of my league if I had to write in Spanish or French.
Now imagine that you don’t spend a lot of time writing English. All of a sudden, you need to put together a paper in the language. Writing well in one’s mother tongue is hard; finding the right turns of phrase in a foreign language is even harder, especially when there are strict page limits, and your field does not have very rigid structures for academic papers.r What do you do?
And, most importantly, as a reviewer, how do you help authors who struggle with their writing?
My own strategy represents a trade-off between time required to review the papers I’ve taken on and diligence. I point out major errors, in particular where terms are used in ways that prevent an English speaker from understanding what is meant, but I let most of the small things slide, in particular when my verdict is “revise and resubmit”. I then provide detailed feedback on the resubmission.
Ideally, journals would have mentors or specialised, paid editors that can help people who struggle with writing English; in the absence of such resources, I often recommend that authors have their papers proofread by a native speaker of English. I know that this can come across as condescending, especially if the authors have worked very hard to write an acceptable paper.
So, what can we do to address this problem as a community? Turning a blind eye to small instances of plagiarism? There are a couple of other options that are relatively inexpensive
- Develop clear language standards, and enforce them when reviewers whose native language is English expect literary masterpieces.
- Put together links on field-specific English for Academic Purposes that authors can access.
- Provide guidance on rewording results and findings from papers for the literature review that helps authors negotiate the line between reporting and plagiarism
- Provide reviewers with the option of submitting annotated PDFs of the paper together with their review – it’s very cumbersome to make a long list of page, line, and paragraph numbers, copy the bad wording, and type out the correction, especially if line numbers do not line up properly with the lines in the text (or when there are no line numbers at all)
What are your suggestions?
May 15, 2012
In circumstances where it can take clinicians and therapists a long time to reach the patients who need them, or where patients need to travel long distances in order to see a specialist, telemedicine comes into its own. Telecardiology or teleradiology allow specialists to receive and assess data from remote locations, giving instant feedback if the communication infrastructure is in place. Thus, telemedicine bridges a gap in knowledge and expertise.
This can be extended to medical education, as Mark Barr from Intel showed at the recent Med-e-Tel conference in Luxemburg. In his presentation, he drew up a knowledge pyramid, where specialists have the highest level of knowledge, followed by generalists, medical nurses, and health workers. Medical education was one of the ways of bridging this knowledge gap.
However, I think that there are really two knowledge pyramids – one of medical knowledge required to help the patient, and one of implementation knowledge required to make sure the patient can get the help they need and implement the required measures. Fitting them both together leads to a continuum where one source of knowledge increases and the other decreases.
For example, if the specialist recommends regular exercise, such as brisk walking for thirty minutes a day, the local health worker can tell people about good routes and point them to local walking groups, if they exist. If the specialist recommends a healthier diet, the health worker can help with suggestions of cheap, nutritious meals, local sources of good ingredients, or cookery classes.
The specialists do not need the local knowledge just like the health workers don’t need the specialist knowledge, but both ends of the continuum need to work together for best results. A top-down conceptualisation of telemedicine, where education just flows along the medical knowledge path, but not back along the implementation one, is – to my mind at least – deeply flawed.
May 8, 2012
There are many low and high tech solutions for tracking one’s moods and feelings, from the humble notebook to the shiny app, from detailed, free-form diaries to ticking a couple of boxes on a form. Many people track their mood informally using social media, letting their online support network know how they feel through blog posts, tweets, and status updates.
Moodscope is a web application that allows users to log and share their mood. Users can rate how you are feeling using 20 mood and emotion adjectives, such as scared, guilty, ashamed, proud, or alert. For each adjective, users need to specify whether they feel like this very slightly or not at all (0), a little (1), quite a bit (2), and extremely (3). This can be done as often as the user wishes.
The interface design is &ldots; unique. For each adjective, the user sees a card. On one side, there are the numbers 3 and 2, on the other side 0 and 1. The user then turns and spins the card until the number that corresponds to the current intensity of that particular emotion is on the top. To log that number, users click on the verbal description. At the end, scores are converted into percentages, where a high percentage reflects good mood, and a low percentage bad mood. After the first three recorded values, a summary is added that describes trends verbally and comments on how frequently Moodscope is used. Daily emails with short, motivating texts serve as reminders. These emails are sent whether or not the user has already completed a mood log that day.
The social aspect is very discreet. Users can nominate buddies who will be sent their Moodscope scores. Buddies need to confirm that they are willing to receive the email updates. All the buddy sees is the current percentage, but there’s an opportunity to discuss scores privately on the Moodscope web site.
One User’s Experience
I started using Moodscope a few weeks ago; I have only used the free version, and my comments may not apply to the paid version.
What first struck me was the pared down functionality. You confirm today’s date, and then it’s straight onto rating the twenty adjectives. Operating the rating interface is slow and cumbersome. You need good eye sight to make out the letters against the background (cards are either reddish or blueish), and the writing can be difficult to read. You also need to plan the flips and turns required in order to get the number you need to enter. This is a strain, and it definitely deters me from logging my mood more than once a day. I am not sure how well a person would do who is easily discouraged, or whose mood makes focusing very difficult.
The summary is not very helpful. As I am logging, my mood shows large swings, but that mostly depends on variation in my mood throughout the day. Moodscope as it stands misses this zig-zag, because the summary only compares my scores to the all time maximum, all time minimum, and average. If there is a large dip, the summary suggests that specific events may have happened to change my mood, when this is nothing but my normal fluctuation.
Fortunately, Moodscope also has a great graphic display.
The graphical display of the change in mood over time is very useful – it allows you to zoom in and out, clearly shows the range of mood that has been registered over time, and can be adjusted to show shorter or longer time periods.
An important limitation is that Moodscope – for good reason – does not ask people whether they are thinking about suicide or self-harm. If it did, there would be unpleasant implications. First of all, it places unreasonable strain on the buddies who receive the reports. What if people indicate that they are ready to kill themselves, but this is not reported to the buddies? And if it is reported to the buddies, what if they discover the warning after the fact? However, this sensible limitation means that Moodscope can miss a significant improvement in cases where people feel still low, but no longer suicidal.
The main reason I would keep using Moodscope is the buddy function. It means that I can let others know how I am feeling through the relatively private medium of email. Otherwise, the interface would be too much of an annoyance, as registering one answer can mean up to three clicks, which needs to be done twenty times. There are thousands of applications diligently monitoring all public social media activity. Somehow, mood is a bit more private than that – it’s good that there’s an application which keeps it private.
May 3, 2012
On Twitter, Rebekah Higgit reacted to David Willetts’ speech on Open Access to research results by asking – well, what about arts and humanities?
In this short post, I would like to outline a couple of practices that I am familiar with from linguistics and psychology that do not require publisher action, but that can be implemented today, by any researcher, unless they’ve signed a particularly stifling publication contract.
- Directly linking to copies of their work on their own web page. This is the most basic version.
- Circulating drafts: Plenty of papers do the rounds in draft form before they are even submitted for peer review. This happened to a paper I worked on which ended up being published in the prestigious journal Language. If you look at the Google Scholar citation page, quite a few citations predate the actual publication. These come from circulated previous drafts.
- E-mail request systems: Many groups working in psychology have an ingenious mechanism for getting around publisher’s restrictions. They set up a central email address whereby you can request a paper, and either a bot or a person monitoring that address will send you a version, either a pre-print as submitted to the publisher for typesetting, or the Authors’ Copy of the typeset version.
- Pure online journals: I’m aware of a few well-regarded online journals in the humanities and social sciences, such as The Qualitative Report and Forum: Qualitative Research.
So, there are a number of ways of enabling Open Access now – let’s hope Willetts’ team manages to negotiate a model that allows more genuine open access for all branches of learning.
May 2, 2012
Last week, New Statesman journalist Martha Gill published a column in the New Statesman’s Current Account blog about the way in which our own biases affect how we process information. She based her argument not just on her own observations, but on a scientific study.
Unfortunately the actual results of the study are not nearly as neat as Gill would like them to be.
Before we start, let me be clear: Studies can be distorted at any step from the reseachers’ fertile minds to the page or blog post, with the most common culprit being (misreported) press releases. I don’t know where the paper Gill cites got twisted beyond its results; I merely based my analysis on the way she reported the study.
Gill versus Nyhan/Reifler
A study published in the journal Political Behaviour shows just how reluctant people are to engage with facts that don’t support their world-view.
Notice that the reference is not given; for fact fans, it’s Nyhan, Brendan and Reifler, Jason (2010): When Corrections Fail: The Persistence of Political Misconceptions. Polit Behav (2010) 32:303–330. The paper is freely available online, as far as I can see. Ben Goldacre explains much better than me why this matters: It allows readers to check whether conclusions were reported correctly.
In the experiment, conducted in 2005, participants were given fake news stories.
There were two experiments, one in 2005 and one in 2006, that consisted of a total of four studies. This is very important – we will see later why.
These news stories were embedded with false facts: that tax cuts under the Bush administration increased government revenues, that weapons of mass destruction had been found in Iraq and that Bush had banned stem-cell research (he only limited some government funding).
The first experiment in autumn 2005 looked at correcting people’s impression that Iraq had weapons of mass destruction; the second in spring 2006 repeated the earlier study, and repeated it with two more topics, tax cuts and stem cell research. In the first experiment, the story with the “false facts” came from Associated Press, in the second experiment, half the participants were shown stories that purported to be from the New York Times, a notoriously liberal paper, and half saw exactly the same story, but this time, it supposedly came from Fox News, a notoriously conservative news source.
After each statement, the researchers put in an unambiguous correction – and then tested the participants to see if they picked this up.
The supposedly unambiguous correction was in fact a paragraph in the same story that reported findings from a relatively objective source which contradicted the key statement in the first paragraphs. So what we have here is not statement / counterstatement, but rather a classic “he said/she said” structure, where journalists present both views.
Here’s the original text from Experiment 1, together with the correction.
Wilkes-Barre, PA, October 7, 2004 (AP)—President Bush delivered a hard-hitting speech here today that made his strategy for the remainder of the campaign crystal clear: a rousing, no-retreat defense of the Iraq war. Bush maintained Wednesday that the war in Iraq was the right thing to do and that Iraq stood out as a place where terrorists might get weapons of mass destruction. ‘‘There was a risk, a real risk, that Saddam Hussein would pass weapons or materials or information to terrorist networks, and in the world after September the 11th, that was a risk we could not afford to take,’’ Bush said.
While Bush was making campaign stops in Pennsylvania, the Central Intelligence Agency released a report that concludes that Saddam Hussein did not possess stockpiles of illicit weapons at the time of the U.S. invasion in March 2003, nor was any program to produce them under way at the time. The report, authored by Charles Duelfer, who advises the director of central intelligence on Iraqi weapons, says Saddam made a decision sometime in the 1990s to destroy known stockpiles of chemical weapons. Duelfer also said that inspectors destroyed the nuclear program sometime after 1991.
The President travels to Ohio tomorrow for more campaign stops.
(Nyhan & Reifler, 2010, p. 324f.)
Can you see how easy it is to reframe this as a piece that just presents two different points of view? Your evaluation of the correction will depend largely on your view of the CIA, Big Government, and bureaucrats who write reports.
They didn’t. Participants who identified themselves as liberal ignored the correction on stem-cell regulations and continued to believe Bush had issued a total ban. Conservatives not only ignored the corrections on Iraq and the tax cuts but clung even more tenaciously to the false information. Facts had made things even worse.
Well, what actually happened?
First of all, this was never about people actually changing their opinion; the researchers are clear that this is future work. Instead, this was a between-subjects design. Half the participants read the story with the correction before they answered the question, half read the story without the correction.
The question participants answered was designed to measure shift in opinion:
Question to participants:
Immediately before the U.S. invasion, Iraq had an active weapons of mass
destruction program, the ability to produce these weapons, and large stockpiles of WMD, but Saddam Hussein was able to hide or destroy these weapons right before U.S. forces arrived.
- Strongly disagree 
- Somewhat disagree 
- Neither agree nor disagree 
- Somewhat agree 
- Strongly agree 
(Nyhan & Reifler, 2010, p. 325)
The authors looked at four predictors of opinion: whether participants had seen the correction, what their ideology was (on a scale from liberal to conservative), how much they knew about politics, and whether their ideology affected their reaction to the correction. When they used all of these predictors to model participants’ answer to the question, by far the largest effect was political knowledge. In the first study, they also found a clear effect of ideology. The correction was more likely to work the more liberal participants were; it backfired for conservatives. Conservatives who read the corrected text were more likely to believe that Iraq indeed had Weapons of Mass Destruction than Conservatives who didn’t.
In the second experiment in Spring 2006, where they looked at the Weapons of Mass Destruction issue again, that backfire effect for conservatives was gone. The only way they could replicate it was to look at a small subset of their participants who had stated that Iraq was most important for them, but that’s a post-hoc analysis (in other words, fishing for results). There are many possible explanations for this, but the authors argue that the main reason was the change in public opinion between Autumn 2005 and Spring 2006. So, far from showing that journalists are powerless, the results actually suggest that persistent corrections across a range of media might be far more powerful than a single story in a point/counterpoint scenario.
When the researchers used a text about tax cuts, however, they saw the same “backfire effect” in conservatives that they had observed in the Autumn 2005 study. In effect, strong conservatives who read the text with the correction were more likely to believe the incorrect fact than strong conservatives who didn’t see the correction. When Nyhan and Reifer looked at liberals’ reactions to a text about stem cell effects, there was no such backfire effect; liberals who read the correction were less likely to accept the correction, but they didn’t show an increased conviction that the opposite was true.
What about people who are neither strongly liberal or strongly conservative? Well, in all of the four studies Nyhan and Reifer conducted, the effect of ideology was gradual, i.e. Centrists were far less likely to be influenced by ideology than either people with strong liberal or strong conservative views.
(By the way, all participants in this study were undergraduates at a Catholic university; the researchers say that their study needs to be repeated with a more representative sample of the population.)
So, What Are We To Make Of This?
People’s perceptions can change, but they don’t change based on reading a single contradictory story, even if it comes from supposedly trustworthy sources. (In fact, in the second experiment, the purported source of the story (New York Times vs Fox News) did not make any difference to the results at all).
What we need to do is understand why people persist in their beliefs despite contrary evidence. In particular, we need to look at
- how much people already know (which explained a lot of the variation in the data)
- how persistent their beliefs are
- what motivates their reasoning
So what can journalists do? Quite a bit, as it turns out.
First of all, once a belief has been formed, it tends to persevere. So, those journalists and news media who break stories have a unique opportunity to affect people’s beliefs about the issues they report, because first impressions are likely to stick (c.f. Ross and Lepper’s (1980) work on belief perseverance, cited after Nyhan & Reifler 2010).
Secondly, and most importantly, if misinformations are persisting, keep correcting them, and seek as many allies as possible to spread the correct information. People who are ‘‘confronted with information of sufﬁcient quantity or clarity… should eventually acquiesce to a preference-inconsistent conclusion.’’ (Ditto & Lopez, 1992, p. 570).
So if you are a journalist or an activist, and Gill’s column discouraged you, take heart. Change happens, but it takes patience and persistence. Keep going!
Ditto, P. H., & Lopez, D. F. (1992). Motivated skepticism: use of differential decision criteria for preferred and nonpreferred conclusions. Journal of Personality and Social Psychology, 63(4), 568–584.
Ross, L., & Lepper, M. R. (1980). The perserverance of beliefs: Empirical and normative considerations. In R. A. Shweder (Ed.), Fallible judgment in behavioral research: New directions for methodology of social and behavioral science (Vol. 4, pp. 17–36). San Francisco: Jossey-Bass.55.951679 -3.198398
April 19, 2012
Both Jo Brodie and Alice Bell have questioned the wider implications of the “Academic Spring” of open access research papers. Providing easy access to research papers is all well and good, but what about helping people make sense of them?
Jo raises an issue that is particularly important, patients and informal carers researching their own condition on the internet. How are they going to cope with medical journal articles, which are written to tight word counts using rigid structures and highly technical vocabulary? Even doctors need training in those skills!
Alice and Jo suggest that journal papers themselves should be written in a more accessible style. While that would be ideal, there is a fundamental problem with this from a communication theoretic perspective. Text structures, conventions of language use, and technical vocabulary emerge because they allow people who share similar background knowledge to communicate efficiently. Sometimes, the most efficient code is graphical; sometimes, it is mathematical.
To be clear, I am not defending badly written papers. Good, clear, engaging writing is still very important, as anybody who has reviewed a badly written paper at 10pm in a desperate bid to make a review deadline will be able to attest. Good writers use technical terms to make their point succinctly, to indicate their theoretical or methodological background, and to link into a wider body of work on that area.
While it is often possible to explain relevant concepts to lay people, this requires changes to the structure of the text, to ensure concepts are properly defined and anchored before they are used. Another problem is making a judgement call about the level of detail that needs to be explained. Let’s take significance testing. How do we communicate what a p-value of 0.05 actually means in practice? And are p-values the right thing to report? What other background should be given?
Fortunately, there is a way out of this conundrum which should be familiar to most researchers who have written or reported on a research grant – the lay summary. Lay summaries are not an easy way out. In order to write a good summary, you need to boil down your research to its essential goal and its main results. This requires a deep understanding of one’s own results. What have we done? What does it actually mean? What more do we need to know?
Traditionally, abstracts and keywords were deemed to be enough to help readers situate the research reported in a paper. Many journals go even further than this. Elsevier journals encourage graphical abstracts and one-sentence research highlights. Many medical journals require a box that summarises the main contribution of the paper to the state-of-the-art; some even use prompts to help authors structure such a text. Some journals have (or had) English abstracts translated into Spanish, French, or German, to help researchers for whom English is not their native language.
From these traditions and innovations, it is but a small step to short summaries written specifically for a non-specialist audience. Funders have shown the way in developing a standard structure, issuing writing guidelines, and providing examples. Unfortunately, what is still missing is a culture of good editing, which costs time and money, but is crucial, especially for lay summaries. The SPARC initiative led the way here, with professionally crafted texts (here’s an example of a project on auditory reminders that I led), but so far, none of the funders I have written summaries for have followed that shining example.
But who then helps non-specialist readers to critically assess what they are reading? Here is where science blogging comes into its own and where science communicators like Ben Goldacre, Petra Boynton, Ed Yong and Kate Clancy shine, to name but a tiny handful. Bewildered by the sheer variety of science blogs? Try Science Seeker or Research Blogging, from there, go to the big and small networks. By now, there is a critical mass of science bloggers who have been honing their craft for years. Add to that well-crafted lay summaries, and we might be onto a science communication winner.
Edited to add (April 20): Stuart Cantrill alerted me to the fact that graphical abstracts have a long history in chemistry. Here is a link to an editorial on graphical abstracts that he has written for Nature Chemistry (requires registration).
March 27, 2012
One of the three projects I currently work on is the MultiMemoHome project. MultiMemoHome is a collaboration between the GIST Human Computer Interaction Group, University of Glasgow, the Centre for Speech Technology Research (CSTR), University of Edinburgh (where I work), and audiologists at Queen Margaret University, Edinburgh. We work with a charity that supports people who are Deaf and hard of hearing, Hearing Concern, and a telecare technology company, Halliday James.
The aim of MultiMemoHome is to create guidelines for designing home care reminders that are
- effective: It’s easy to understand the message
- adaptable: Reminders can be personalised to match people’s information needs.
- accessible: The reminder is easy to perceive, even for people with mild to moderate sensory impairments.
- acceptable: Listeners don’t get tired of all the reminders, and reminders are pleasant to see, hear, feel, or smell.
In the project, we work with four modalities, hearing, touch, vision, and olfaction (smell). At CSTR, we’re looking at auditory reminders – messages that can be heard. Our aim is to design reminders so that they are easy to understand, even if your hearing is not what it used to be, and even if there is a lot of background noise. Such reminders could be anything from a spoken message to a short tune.
Our work focuses on two questions:
- How well can people understand and learn different types of auditory reminders?
- To what extent do auditory reminders distract people from what they are doing?
To address these question, we are running a series of three studies. In each study, we ask people to act on auditory reminders that they hear while they are doing something else – memorising a sequence of numbers or words, reading text, or walking. We’re particularly interested in walking because falls are a big problem for older people living on their own, and we would not want a reminder to startle a person so much that they trip and fall.
As part of the first research question, Karl Isaac, the Edinburgh project PhD student, is looking at the effect of different types of background noise on the intelligibility of computer-generated messages. Karl is about to extend his work to distortions that can make speech more difficult to understand, such as reverberation due to room acoustics.
If that sounds interesting, and you’d like to get in touch with us about our research, the web site has contact details We also conducted an extensive survey that we used to scope our research plans; it can be found here and feedback is still welcome – no contribution is wasted!
March 20, 2012
A recent blog post by Puffles’ bestest buddy motivated me to put down my own principles of online social behaviour – mostly because Bestest Buddy’s approach is very similar to my own. I’ve also been thinking about these issues after reading a recent post by pme200. (If you are on Twitter, you should follow both.) I have decided to publish these reflections on my work blog because I have been thinking about using social media for research, and I use social media to promote good scholarship and science (in addition to my own work, which may or may not be up to scratch …)
I am interested in how people approach the koan of life – What is a good life? What is our place in this world? How can I muddle through from day to day? The best way of finding that out is to listen carefully. I almost wrote “non-judgmentally”, but of course I have principles I live by, and I have opinions on what I see on- and offline. Still, I find the best way to keep an open mind is to communicate with people with many different views and opinions. I am also aware that many of the people who upset me the most are people who openly display personality traits and attitudes that I am afraid of seeing in myself. In other words, when I see my own shadow, when I see what I could be if I don’t watch out, I recoil.
I interact with people who are open atheists and people who are creationists, with ardent alt-med skeptics and people who happily use chiropractic and naturopathic medicine, with socialists and conservative libertarians, and I enjoy this mix. For the record, I am deeply religious, with liberal Catholic / Quaker views, strongly favour a good social security net, and support evidence-based medicine, with more time for practitioners and patients to establish a strong personal rapport.
When I express my opinions online, I always try not to cause undue offence. This is relatively easy because my audience is the diverse group of people I talk to, and I do my best to keep the people who might be offended in mind when deciding what to post and how to phrase it. Sometimes, people tell me that they are put off by something I post. This usually happens when I haven’t quite thought the post/tweet/status update through properly, and did not frame it well. We talk about what it was that offended them, and I can take their point on board, understand their reactions, and sharpen my own thinking. A few times, this approach has led me to change strongly held views, and I am always happy to apologise for offence I caused inadvertently.
I also keep a relatively tightly locked online space under a pseudonym which I have been using for more than a decade in various forms which I use for support, more personal reflections (AKA relentless whining) and to stay in touch with close friends. My rules of behaviour are similar for my real name identity and for my pseudonym, but under my pseudonym, I am both a bit more open about my own views, and a bit more sensitive to my online friends’ views than I am under my real name.
I greatly value politeness and manners. I’ve been extending my comfort zone recently because I started posting on a forum where the overall tone is rougher than what I am used to.
As a result of this policy, which I’ve developed through trial and error during my 18 years online (I got my first e-mail address in 1994, fact fans), I’ve had relatively little trouble with trolls and internet hate. This is good, because my single biggest weakness is that I am extremely sensitive to trolling. In fact, I am so sensitive to trolls that I am not just suspicious of people who have trolled me in the past – I am also suspicious of people who are friends with said trolls. I’ve observed that trolls tend to hang out with other who have similar manners (or lack thereof), I am wary of trolls getting their bile green warty hands on private information, and seeing the names of certain trolls retweeted into my timeline gives me a small nasty jolt (see above, single biggest weakness).
I should add that my definition of troll is somebody who has trolled others online, irrespective of their offline or other online behaviour. In fact, it is perfectly possible that one and the same person behaves like a vicious troll towards some people, and like a wonderful, steadfast friend towards others. Everybody is capable of trolling, harassment, and vile behaviour; trolls are people who act this out in public.
My other weakness (paradoxically) is that I am actually quite an irascible person, and when I’m particularly low and vulnerable, I’m not quite as polite and well-reasoned as I should be. I’ve been reined in a couple of times over the past decade by friends when I overshot the mark online, and I am very grateful to them (you know who you are!).
Having followed the anonymity / pseudonymity discussion for a while, and having both an open and a pseudonymous identity, I have strong views on the matter. For example, I like the idea that pseudonym identities aren’t judged based on their position in life, but on the quality of their output. The much-missed Revere epidemiology blogger(s) wrote a great post about that idea. I don’t think attacking people from behind a pseudonym is cowardly – if that pseudonym is the person’s online identity. It would be cowardly to be sweetness and light under one’s own name and a total pain in the neck under another handle.
At the end of the day, be consistent and be true to yourself – as long as you don’t commit a crime in the process, or cause substantial damage and pain to another person, that’s all that matters, and you will find people online who are willing to interact with you and accept you the way you are. That’s the horror and the beauty of the Internet.