Research Writing Pointers

Research Interviewing: Some Pointers

Interviewing is one of the most powerful tools we have for answering research questions in the counselling and psychotherapy field—indeed, any field. ‘Interviewing’ is essentially just a technical term for ‘talking to someone’—and it’s a talking that can have a great deal of depth, nuance, and complexity. What better way of finding out the answers to your questions than by going straight to the source: people, real people, and asking them what they experience or what they think. It’s not the only way of answering research questions, and it’s not always the most suitable; but if you’ve got a question about how a particular group of people experience or perceive something, it’s one of the most effective and widely used tools.

Types of interviews

Interviews can be structured, semi-structured, or unstructured. That’s essentially the amount of a priori ordering you give to an interview: how much do you run it along pre-determined lines, with the same questions, asked in the same way, to each participant? Of course, there’s no right or wrong here—it all depends on what kind of questions you are asking. Structured interviews tend to be best for smaller q (i.e., more quantitative-like) research. For instance, if you wanted to compare how much gay and bisexual men found gay-affirmative therapies helpful, you might want to use a structured interview format so that you had clear, comparable answers to analyse. If, on the other hand, you are doing larger Q research (i.e., more in-depth, experiential, explorative) an unstructured format may be better so you can really go ‘with the flow’ in each interview and explore as deeply as possible their responses. In the middle is a semi-structured approach, which is often a good compromise, and means that you are asking every respondent similar questions, but you have flexibility to go into depth and follow up particular areas without worrying too much about ‘sticking to the script’.

Interviews can also be scheduled, unscheduled, or semi-scheduled, which refers to whether questions are asked in the same order or not. Scheduled tends to go with structured, and is more suited to research questions where its important to reduce biases. Unscheduled, by contrast, gives you the flexibility to go into depth in an interview, as and where the respondent seems to be opening areas up. Again, semi-scheduled can be a good compromise, meaning there’s a provisional order to your interview, but scope to go ‘off track’ if it feels like moving the order around will be productive.

As suggested above, a semi-structured, semi-scheduled approach is a good option—and commonly used—alongside a wide variety of qualitative counselling and psychotherapy research methods, including interpretative phenomenological analysis (IPA), grounded theory, and thematic analysis. However, if you are wanting to conduct a very large Q grounded theory or thematic analysis study, a wholly unstructured, unscheduled approach may be best.

Interview Schedule

However you do your interview, even with the most unstructured of approaches, it’s generally good to prepare a schedule or protocol of how the interview is going to go. This is absolutely essential if others are going to be doing the interview as well as you, so that they know what to do and your interview is consistent across participants.

In one of my first research studies, on therapists’ experiences of relational depth, I developed a very basic interview schedule, which consisted just a few questions and prompts:

  • Could you tell me about a time when you really felt that you met with a client at a level of relational depth?

    • What was that experience like?

    • What did you take from that experience?

    • What do you think your client took from that experience?

    • How did that experience come about?

  • What facilitates your ability to meet clients at a level of relational depth?

  • How important do you think that meeting clients at that level of depth is?

An example of a much more extensive interview schedule, for our study of clients’ experiences of working with preferences, can be found here. This was designed for use by myself and another interviewer; and with a team of very experienced international researchers discussing, drafting, and finalising it. As an interview schedule, however, the heart of it was still the questions to be asked. In this case, the main question was:

  • What, if any, preferences did you have when you came into therapy?

And then, for each preference, we asked:

  • Where do you think this preference might have come from?

  • How, if at all, did this preference change over the course of therapy?

  • How important was it to you that this preference was (a) elicited/discussed, and (b) accommodated in the therapy?

When writing a schedule, it may be useful to note follow-up prompts for different questions: potential responses you can use after the main question to help the participants go deeper into the exploration. For instance, for the question:

  • Where do you think this preference might have come from?

You might then prepare prompts like:

  • Was it related at all to your earlier life?

  • Have you had therapy before? Was it based on those experiences?

  • Are there any other origins you can think of?

On our schedule for the preference study, you’ll see that we also included:

  • Notes on the general approach/style to be taken in the interview

  • Materials to be taken to the interview and other general instructions

  • Approximate wording for the introduction to be given to participants

  • Additional questions

In this case, there’s also some guidance on questions that were asked outside of the scope of the main study.

Generally, there’s no harm having as much detail as possible in your schedule, even if it’s about engaging with your participants in an entirely unstructured and unscheduled way. Write it down—it’s a great reminder for you about what to do, something you can draw on when you write up your Methods section and, if you’re writing a thesis, a great resource to put in your Appendices: showing your examiners that you were really clear, consistent, and methodical in what you did.

Piloting

Before finalising your schedule, and definitely before running your interviews, try the interview out on someone—ideally a few people. And who better to start with than yourself. As well as giving you a real sense of what it’s like being interviewed with this schedule (and, perhaps, what’s working and what needs changing), it can help you unpack your biases and expectations with respect to the research question (again, something you may then want to write up and include as an appendix). Even if you aren’t part of the target group for the research study, you can role play it. Then try it out as an interviewer—perhaps with friends, colleagues, or peers—to get a sense of how it feels running the interview (bear in mind you may need ethical approval before doing this). If they are not part of your target group, again, they can always role play, though you will probably want to also pilot the interview with someone who is part of your target group before moving in to the main phase of interviewing. And remember, when you write up, you can describe this piloting process, showing your care and methodicalness in developing and testing out your interview schedule.

In fact, I generally encouraged my research supervisees to transcribe their pilot interviews, and even do some initial coding or analysis on them. That’s not suitable to all methods (particularly if you’re keen to go into each subsequent interview with an entirely open and ‘unbiased’ mindset), but it’s a great way of checking whether you’re data is making sense: that is, whether you’re actually getting meaningful answers to the questions that you’re asking. At the very least, ask yourself: ‘What have I learnt from this/these interviews (including the one with yourself) to the question(s) that I am asking?’ If the answer is ‘nothing’ or ‘I’m really not sure’, you may want to revisit your interview schedule or the focus for your project. Ideally, you’re coming away thinking, ‘Wow, those were some great initial answers to my questions… I’m hungry to learn more.’

Can you include data from your pilot interviews in your main analysis? Ideally, with most methods, not—formally, it’s an activity where you are preparing for the actual method you are going to use. But, as above, it’s great to write up your pilot interviews in your Methods section, and particularly what you have learnt from them. But some approaches, particularly larger Q ones (where there’s less focus on consistency or unbiasedness across interviews), may see less problems folding in your pilot interviews into your main data. If you’ve got really rich and meaningful data from them, why not use it?

Dialogue

Except for very structured, ultra-small q interviews, one of my main pointers for conducting interviews in counselling and psychotherapy research would be to see them as dialogues—with the aim of finding answers to the question(s) you are asking. Too often, particularly with trainees, there is a tendency to conduct interviews a bit like TV’s Question Time: with lots of questions, one after the other—and the interviewer’s nose in their interview schedule—without enough time, space, or attention to really explore the participants’ experiencing or perceptions in depth. I’m sure this often comes out of the interviewer’s anxiety: a worry about what to say next, and a concern to keep things going. But an overly-scheduled interview, with lots of quick fire questions, generally doesn’t allow the richness and depth of a participant’s experiences to emerge. People aren’t always able to answer well the first time around—to say what they really mean—so the interviewer needs empathic reflections, follow-up questions, prompts, paraphrases, summaries (many of the skills you may have learnt as part of a counselling training) to help the participant unpack what has really gone on for them. Below is an example of such dialogue from our study of clients’ experiences of preference work. The participant has already said that they came into therapy with a preference for ‘structure’ but, as an interviewer, I’m using reflections and further questions to try and understand more about what that means.

Interviewer: What would you say your main preferences were when you came into the therapy? You’ve already said about– there was one about structure, there– say a bit more about that.

Participant: ….So like you say, structure was important to me. I didn’t– I’ve never done anything like this before, so it’s completely new to me. I didn’t know what I should be talking about, or how to achieve those goals. So it was more kind of guidance, and what can we do to– that’s going to [pause] […] less anxious, um, I don’t know. I don’t know. I’m completely lost about how to get to those goals. So my structure would be what– what can we do to bring about those changes, how do we work towards that, what– um, so that at the end of the session, you know, I’d think– or I’d come back the next week and say, ‘Have you felt less anxious?’ and my initial reaction was that, ‘Well no, I haven’t felt less anxious, because why would I, nothing’s changed. I’ve done nothing to change it. I don’t know what to do to change it. Um… I’ve had no guidance’, so, yes, my- my answer to the question would be the same, it’s […] unless I’m doing something different.

Interviewer: So you really wanted to be… guided, and have something to do, say in between sessions?

Participant: Yeah, that would have been nice. Yeah. You know, I’ve never done this before, so I didn’t have any expectations. It was just, ‘Let’s see where it goes.’

Interviewer: Yeah, but you wanted– What you’re saying– Did you have that, like when you started therapy, were you thinking, ‘I’d really like some structure,’ or was it something that kind of emerged as the therapy continued?

So we don’t just move on from ‘a preference for structure’ once the participant has said it. Rather, we circle around it, try and go deeper, explore what the participant really means by it and how it links to other responses. It’s a process, all the time, of deepening understanding—again, just as you might do in therapy if you were trying to help a client make sense of particular experiences.

You can see that, in this last question, I’m trying to understand in more depth how this preference emerged. What’s going on in my head is, ‘I really want to understand what the participant wanted in therapy and why?’ That, those core research questions, are my focus, and I’m trying to engage with the participant in a range of different ways so that we can answer those questions. So, at those points, I’m not really thinking about my interview schedule. It’s something I’ll come back to, and I’ll try and make sure I’ve asked all the questions I’ve wanted to; but, at those times, it’s the overall objective that is the orienting point to the interview. I genuinely, deeply, want to discover—with the participant—what the answers are to my research questions. Indeed, I sometimes think that the best interviews are when the interviewer has just one main question, and riffs around it, with the interviewee, in a variety of ways to answer it as best they can.

Interpret along the way

What you can see in the interview above, and my second key pointer for most kinds of medium-to-large Q interviewing, is to try and make sense of what the participant is saying as you go along. Far better than trying to interpret what the participant was saying once you’ve got your transcript, do it as you go along with the participant. Because you’ve got them there to help you, and guide you, and tell you what it is that they actually mean. For instance, in the example above, they’d said that they wanted ‘structure’, but what did this actually mean, and were they saying they wanted guidance or something else? I wasn’t sure, and I asked.

Lots of times in interviews, you’ll be sitting there thinking, ‘I’m not really sure what this person is trying to tell me.’ That’s nothing to be ashamed of—it’s a really valuable indicator to you that you need to be asking more. You can even say to a participant, gently and kindly, something like, ‘Sorry, I didn’t quite get that, can you explain it to me?’ or ‘I think what you’re saying is …, Is that right?’ If you leave an interview really confused by what a participant has told you, chances are you’ll struggle to make much sense of it once it’s transcribed. And, of course, participants won’t always be able to make sense of things or tell you what they really mean. But to the extent that participants can explain and describe their experiences in depth, use the interview—itself—as a way of unpacking all that richness and meaning. In my own experience, I’ve been amazed by how much research participants can actually make sense of their own ‘material’, themselves—given time, space, and a dialogical sounding board to help articulate and explicate on what they are saying. And that way, when you leave the interview, a lot of the hard work of the analysis has already been done.

Just related to this, a point my colleague Gina Di Malta raised, is that it’s important to try and stick to the participant’s own words as much as possible, rather than introducing your own terms and ideas (again, a bit like therapy). You want to interpret, but you want that interpretation to be one of ‘unpacking’ the participant’s meanings, not introducing terms or conceptualisations from outside. So a question like:

  • You said you wanted ‘structure’, can you say more about what that means to you?

Is probably better than one like:

  • By structure, do you mean having an agenda for the sessions?

Given the power differential in the interviewing situation, a ‘yes’ to the second question really isn’t going to tell you much. Did they really want to have an agenda for the sessions, or were they just ‘deferring’ to you to be polite?

keep to task

Key pointer #3: you’re not the participants’ therapist, and you need to come away from the research interview with answers to the research questions that you’re asking (at least, in the ideal). Something I see a lot in supervisees’ interviewing, particularly when they’re training in a person-centred or relational approach, is that they tend to fall in to a non-directive, therapeutic mode of being: following the participant in whatever the participant wants to talk about. Of course, it’s important to be sensitive, empathic, and non-judgmental when you’re interviewing—and not overly-authoritarian—but participants may well come into the interview with you with things that they want to talk about or tell you about, and what they want to say may not relate in any particular way to what your research questions are asking. A participant, for instance, might want to tell you about difficulties in their marriage when you’re actually focusing on their experiences of a particular service; or they might want to tell you about their experiences of a service when your research project is actually focused on their childhood experiences. Participants have every right to want to tell you what they want to tell you, but it’s both ethically and methodologically important that you help steer the conversation to what you’re research is about. Otherwise, you can end up leaving the interview without any useable material and, more importantly, your participants can go away feeling confused and exposed by a process that was different to what they were invited in to.

So even though participants may take the interview off in all kind of directions (and that’s their right to do so), your role is to gently, sensitively, and compassionately bring them back to the main points of discussion. That doesn’t need to be rude or abrupt: if, for instance, a participant is really going off track, you can gently ask them how it relates to your key research questions, or thank them for their response and go on to the next question. And, in any interview, there’s bound to be material that is less relevant to the researcher’s question(s). In a sense, the issue is less about there being lots of ‘irrelevant’ material, and more about their being insufficient relevant material to be able to answer your research question. Having said that, transcribing and then trying to analyse page after page of material that, actually, isn’t much use to your research can be quite a heart-sink. Again, that’s why it’s really useful to do a few pilot interviews and a bit of an analysis, to make sure that the way your interview is set up is getting you the answers you want.

Keeping an interview on track means that, at times, you may need to be quite directive. For therapists schooled in the idea that directivity is just one step away from satanic worship, adopting a more proactive and leading stance can feel challenging: the natural tendency is just to go with the client/participant and not interrupt, structure, or lead. And, as above, there’s a lot of relational counselling skills that can be really valuable in interviewing work. But interviews do need to be more than just an unstructured, unfocused chat: again, not just for you, but out of an ethical duty to your participants to stick to ‘what you said on the tin’—they didn’t sign up for therapy, so you can start providing it for them.

Below is an excerpt from an interview, from the same study on client preferences, which demonstrates how an interviewer can gently and sensitively focus the interview process. In this case, the participant has been talking about a preference for a therapist who was challenging-but-supportive, but then goes on to talk about a preference for a female therapist. The interviewer positively affirms the expression of the latter preference, but then invites the participant to say more about the original preference so that they can explore it in more detail and systematically:

Interviewer: I guess there are lots of different preferences back on who your therapist is, and what the…

Participant: Oh yes.

Interviewer: …treatment is and…

Participant: Yes, I did prefer women because I just thought I would be more relaxed with that.

Interviewer: That's a good point. That's a really good example of preference, but what you said before, I'll just come back to that because that was really interesting as well. You were saying that you kind of have a preference, so someone who was a bit challenging but also quite supportive. How would you phrase that one?

Fear

Let’s be honest, interviews can be pretty scary to do especially if, like me, you’re natural tendency is towards being quite shy and unconfident. You’re interviewing someone who, pretty much by definition, knows more about something than you, and you may be interviewing people in positions of authority or power. You want to look good: you don’t want to babble away, lose your recording device, or have participants leave the interview thinking, ‘What kind of moron was that!’ Recognising any fears, of course, is the first step in managing them, and talking them through with peers or supervisors can be a really valuable next step. That can help normalise any anxieties, and also help you think about strategies for dealing with them. Ideally, you’re going into the interview feeling pretty calm, so that you can really listen to your participants. You want to have your head in your participants’ worlds—not in your own anxieties. And you also need a degree of confidence to keep your participant to task.

Having said all that, it’s also worth remembering the Yerkes–Dodson law: a bit of arousal isn’t a bad thing, and can help you be focused, attentive, and organised. It also conveys to your participant that you’re taking them and their time seriously: too laid back and your participant might feel a bit taken for granted. And, whatever else, please don’t fall asleep while doing an interview!

Stonewalled

What if you’re interviewing someone who, however, hard you try, doesn’t seem to want to give you much, or just doesn’t seem to be able to give in-depth responses? You’re getting short, one sentence answers; you probe, prompt, try to ‘unpack’ what they’re saying, but it’s just more one sentence responses, ‘I don’t knows’, or even silences. Or perhaps your participant has something that they’re so keen to tell you that you just can’t keep them to task: nothing they are saying—however hard you try—is related to your research questions.

Of course, ethics comes first. To repeat, your participant has the right to come into the interview and do, or not do, what they want (within reason), so if it’s clear that they don’t want to say more or go into more depth, that has to be respected. More than that, it’s important that the participant comes away from the interview feeling that their contribution has been valued, and not that they’ve let you down in some way.

In my own experience and that of my research supervisees, it’s very much the norm that, say out of eight interviews, at least two or three might be more limited in content. So if you’re finding that you’re not getting much from a particular interview—and you’ve tried, as much as would be reasonable and respectful, to invite more, without success—then it’s generally best just to accept that that is how that interview is going to be. In a sense, the quality and depth of a research project is not defined by how many ‘bad’ interviews you have but how many ‘good’ ones there are. And one option, if you are finding that a number of your interviews are quite limited, is simply to do more. Having said that, even where participants give relatively limited responses, the material may still be very helpful: giving an indication of the range of different responses that may be out there. Also, if you’re doing a smaller q study, with some quantification of responses, more limited interviews can still help you develop a sense of how many people experience or see things in a particular way.

Of course, if it’s the case that a large proportion of your respondents can’t give you in-depth responses to your interview questions, then that probably says something important about the way your interview is structured, or what you are trying to do. Again, that’s why it’s so important to pilot your interview schedule and to see whether people—including yourself—can give rich, deep, and interesting answers to the questions you are asking.

Ethics

A few further words on the ethical side of the research interviewing process.

Ethics, as always needs to come first. Interviews can be extremely rewarding for participants, but there are also some risks to be mindful of, particularly if you are exploring highly sensitive topics (for instance, clients’ experiences of childhood difficulties). Some pointers:

  • Make explicit in the information sheet what the research will involve, including potential risks, so that prospective participants know what they are signing up for.

  • Empathy and a non-judgmental acceptance should be de rigueur throughout the interviewing process: to repeat, you want your participants to leave feeling that their experiences and perceptions have been heard and are of value.

  • Pay close attention to maintaining your participants’ confidentiality, particularly if materials are going to be made publicly available (for instance, on a thesis University depository or in a journal). Also, be extremely careful with audio files—it’s a lot easier for participants to be identified if their voices, as well as the content of what they say, is disclosed.

  • Hope for the best but plan for the worst. What happens, for instance, if your participant gets extremely distressed during your interview and says that they are going to kill themselves? Make sure you have strategies in place to respond to such emergencies (for instance, making it clear in your information sheet that you would tell their GP if there was serious risk of harm, and providing links to support services like the Samaritans).

  • Be careful (as above) not to slip into a therapist role. Some of your interview may be therapeutic for the participant, but the focus should remain on answering your research questions rather than healing for the participant, per se.

Read up on it

Finally, as with any method, it’s important to explore the contemporary literature on interviewing, to get a deeper sense of how it can be done, the issues that can emerge, and different tips and strategies for optimising benefit. My own personal bible on interviewing, for many years, was Steinar Kvale’s (1996) InterViews, now in its third edition as Brinkmann and Kvale’s (2014) InterViews: Learning the craft of qualitative research interviewing. The book gives a brilliantly humanistic, relational guide to the interviewing process. Sage also have a series of videos on in-depth interviewing that you may be able to access.

Acknowledgement

Thanks to Gina di Malta, and trainees on the PsychD Counselling Psychology Programme at the University of Roehampton for suggestions and advice.

Photo by Christina@wocintechchat.com on Unsplash

DISCLAIMER

The information, materials, opinions or other content (collectively Content) contained in this blog have been prepared for general information purposes. Whilst I’ve endeavoured to ensure the Content is current and accurate, the Content in this blog is not intended to constitute professional advice and should not be relied on or treated as a substitute for specific advice relevant to particular circumstances. That means that I am not responsible for, nor will be liable for any losses incurred as a result of anyone relying on the Content contained in this blog, on this website, or any external internet sites referenced in or linked in this blog.

Measure Development and Testing Research: Some Pointers

Have you ever had one of those dreams where you’re running towards something, and the faster you go the further away it seems to get? That, to me, is what doing research in the measure development field seems like. Every time I think I have mastered the key methods some bright spark seems to have come up with a new procedure or analysis that is de rigueur for publishing in the field. No mind, I have to say that developing measures has been one of the most satisfying and even exhilarating elements of my research career, however humbling it might be at times. And, indeed, having gone from knowing next to nothing about measure development to creating, or helping to test, some fairly well-used measures (including one with my name on it, the Cooper-Norcross Inventory of Preferences!), I’m pretty confident that it’s a research process that anyone—who’s willing to devote the time—can get involved in.

And, of course, the point of developing and validating measures is not just for the narcissistic glory. It’s research that can help to define phenomena and explore their relationship to other factors and processes. Take racial microaggressions in therapy, for instance. Measures can help us see where these are taking place, what’s leading to them, and help us assess methods for reducing their prevalence. Of course, the downside of measures is that they take complex phenomena and reduce them down to de-contextualised, linear variables. But, in doing so, we can examine—over large, representative samples—how these variables relate to others. Do different ethnic groups, for instance, experience different levels of racial microaggressions in therapy? We could use qualitative methods to interview clients of different ethnicities, but comparing their responses and making conclusions is tricky. Suppose, for instance, of the Afro-Caribbean clients, we had four identifying ‘some’ microaggressions, two ‘none’, and three ‘it depended on the therapist’. Then, for the Asian clients, we had two saying, ‘I wasn’t sure’, three saying ‘no’, and two saying, ‘it was worse in the earlier sessions’. And one Jewish client felt that their therapist made an anti-Semitic comment while one didn’t. So who had more or less? By contrast, if Afro-Caribbean clients have an average rating of 3.2 on our 1 to 5 scale of in-therapy racial microaggressions, and Asian clients have an average rating of 4.2, and our statistical analysis show that the likelihood of this difference being due to chance is less than 1 in a 1,000 (see blog on quantitative analysis), then we can say something much more definitive.

From a pluralistic standpoint, then, measure development research—like all research methods—has a particular value at particular points in time: it all depends on the question(s) that we are asking. And while, as we will see, it tends to be based on positivistic assumptions (that there is a real, underlying reality—which we can get closer to knowing through scientific research), it can also be conducted from a more relativist, social constructionist perspective (that no objective ‘reality’ exists, just our constructions of it).

What is Measure Development and testing Research?

Measure development research, as the name suggests, is the development of ‘measures’, ‘scales’, or ‘instruments’ (also known as the field of psychometrics); and measure testing research is assessing those measures quality. Measure development studies will always involve some degree of measuring testing, but you can have measure testing studies that do not develop or alter the original measure.

A measure can be defined as a means of trying to assess ‘the size, capacity, or quantity of something’: for instance, the extent to which clients experience their therapist as empathic, or therapists’ commitment to a spiritual faith. In this sense (and particularly from a positivist standpoint), we can think of psychological measures as a bit like physical measures, for instance rulers or thermometers: tools for determining what’s out there (like the length of things, or their temperature).

Well known examples of measures in the counselling and psychotherapy field are the CORE-OM (Clinical Outcomes in Routine Evaluation – Outcome Measure), which measures clients’ levels of psychological distress; and the Working Alliance Inventory, which measures the strength of therapist-client collaboration and bond. There’s more information on a range of widely used ‘process’ and ‘outcome’ measures for counselling and psychotherapy here.

Measures generally consist of several ‘items’ combined into a composite score. For instance, on the CORE-OM, two of the 34 items are ‘I have felt terribly alone and isolated’ and ‘I have felt like crying’. Respondents are then asked to score such items on a wide range of different scales—for instance, on the CORE-OM, clients are asked to rate the items from 0 (not at all) to 4 (Most or all of the time)—such that a total score can be calculated. Note, in this way, measures are different from ‘questionnaires’, ‘surveys’, or ‘checklists’ that have lots of different items asking about lots of different things. Indeed, as we will see, the ‘combinability’ of items into one, or a few, scales tends to be a defining feature of measures.

A measure can consist of:

  • One scale. An example is the Relational Depth Frequency Scale, which measures the frequency of experiencing relational depth in therapy.

  • Two or more scales. An example is the Cooper-Norcross Inventory of Preferences, which has scales for ‘client preference for warm support vs focused challenge’, and ‘client preference for past focus vs present focus’.

  • Two or more subscales: meaningful in their own rights, but also summable to make a main scale score. An example is the Strengths and Difficulties Questionnaire for children, which has such subscales as ‘peer problems’ and ‘emotional symptoms’, combining together to make a ‘total difficulties’ score.

Generally, A single scale measure or a subscale will have between about four and 10 items. Less than that and the internal consistency starts to become problematic (see below); more than that and the measure may too long to complete, with items that are redundant.

Measures can be designed for completion by therapists, by clients, or by observers. They can also be nomothetic (where everyone completes the same, standardised items), or idiographic (where people develop their own items, for instance on a Goals Form).

Underlying Principles

Most measure development and testing research is underpinned by a set of principles known as classical test theory. These are fairly positivistic, in that they assume that there are certain dimensions out there in the world (known as latent variables) that exist across all members of the population, and are there independent of our constructions of them. So people’s ‘experiencing of racial microaggressions’ is a real thing, just like people’s temperature or the length of their big toe: it’s an actual, existent thing, and the point of our measure is to try and get as close as possible to accurately assessing it.

You might think, ‘If we want to know about clients’ experiences of racial microaggressions in therapy, why don’t we just ask them the question, “To what extent do you experience racial microaggressions in your therapy?”’ The problem is, from a classical test theory perspective, a respondent’s answer (the ‘observed score’) is going to consist of two components. The first component is going to be the part that genuinely reflects their experiencing of microaggressions (the ‘true score’ on the latent variable). But, then, a second part is going to be determined by various random factors that influence how they answer that specific question (the ‘error’). For instance, perhaps the client doesn’t understand the word ‘microaggressions’, or misunderstands it, so that their responses to this particular item don’t wholly reflect the microaggressions that they have experienced. Here, what we might do is to try and minimise that error by asking the question in a range of different ways—for instance, ‘Did your therapist make you feel bad about your race?’ ‘Did your therapist deny your experiences of racism?’—so that the errors start to even out. And that’s essentially what measure development based on classical test theory is all about: developing measures that have as little error as possible, so that they’re evaluating, as accurately as they can, respondents’ true positioning on the latent variable. No one wants a broken thermometer or a wonky ruler and, likewise, a measure of the experiencing of racial microaggressions in therapy that only reflects error variance isn’t much good.

As you can see, all this is based on very positivist assumptions: a ‘true’, underlying (i.e., latent) reality out there in the world; acting according to laws that are true for us all; and with ‘error’ like an uninvited guest that we’re trying to escort out of the party. Not much room for the existence of unpredictability, chaos, or individual uniqueness; or the idea that ‘reality’ is something we construct according to social mores and traditions. Having said that, adopting classical test theory assumptions, for the purposes of measure development, doesn’t mean you have to be a fully-fledged positivist. From a pragmatic standpoint, for instance, you can see measure development as a means of identifying and assessing something of meaning and importance—but whether or not it is something ‘real’ can be considered a mute point. We know, for instance, that there is something like racial microaggressions that can hurt clients and damage the therapeutic relationship, so we can do our best to find ways of assessing it, while also acknowledging the inherent vagaries of whatever we do. And, perhaps, what we call ‘racial microaggressions’ will change over time and vary across cultures and individuals, but that shouldn’t stop us from trying to get some sort of handle on it, so that we can do our best to find out more and intervene.

Developing a measure

So how do you actually go about developing a measure? It might seem like most measures are developed on the back of the proverbial ‘fag packet’ but, OMG, it is vastly more complicated and time-consuming than that. I worked out that, when Gina di Malta (with myself and Chris Evans) developed the 6-item Relational Depth Frequency Scale, it took something like six years! That’s one year per item.

That’s why, for most of us who have developed measures, the first thing we say to people who want to develop their own measures is to first see if they can use measures that are already out there. That’s unless you really have the time and resources to do the work that’s needed to develop and validate your own measure. Bear in mind, a half-validated measure isn’t much valid at all.

So why does it take so long? To a great extent, it’s because there’s a series of stages that you need to go through, detailed below. These aren’t exact, and every measure development study will do them slightly differently, but the sections below should give you a rough idea of what steps a measure development study will take.

Defining the Latent Variable

Before you develop a measure, you have to know what it is that you are trying to measure. To some extent, this may emerge and evolve through your analysis, but the clearer you are about what you’re looking for, the more likely your measure will be fit for finding it.

‘I’d like to know whether clients feel that they’ve got something out of a session.’ OK, great, but what do we mean by ‘got something out of’? Is this feeling that they’ve learnt something, or finding the session worthwhile, or experiencing some kind of progress in their therapy? ‘Maybe all of those things.’ OK, but feeling like you’ve learnt something from a session may not necessarily correlate with feeling like you’ve made progress. They may seem similar, but perhaps some clients feel there’s a lot they’ve learnt while still coming out of a session feeling stuck and hopeless.

Things that just naturally seem to go together in your mind, then, may not do so in the wider world, and disentangling out what you want to focus on is an important starting point for the measure development work. How do you do that? Read the literature in the area, talk to colleagues, journal, look at dictionaries and encyclopaedias: think around the phenomenon—critically—as much as you can. What you want to identify is one discrete variable, or field, that you can really, clearly define. It could be broader (like ‘the extent to which clients value their sessions’) or narrower (like ‘the extent to which clients feel they have developed insight in their sessions’), but be clear about what it is.

item generation

Once you know what latent variable you want to measure, the next step is to generate items that might be suitable for its assessment. At this stage, don’t worry too much if the items are right or not: brainstorm—generate as many items as you can. In fact, one thing I’ve learnt over the years is that you can never have too many items at this stage, and often you can have too few. Probably around 80% or so of items end up getting discarded through the measure development process, so if you want to end up with a scale of around 5-10 items, you probably want to start with around 25-50 potential ones. Bear in mind that you can always drop items if you get to the end of the measure development process and have too many, but it’s much more difficult to generate new items if you get to the end and find you have too few.

Ideally, you want to do this item generation process in one or more systematic ways, so it is not just the first, ad hoc, items that come into your head. Some strategies for generating items are:

  • Search the literature on the topic. Say we wanted to develop a measure to assess the extent to which adolescent clients feel awkward in therapy (we’re interested in differences in awkwardness across types of therapies, and types of clients). So let’s go to Google Scholar to see what papers there are on young people’s awkwardness in therapy, and also we should check the more established psychological search engines like PsychInfo and Web of Science (if we have access, generally through a university). Supposing, there, we find research where young people say things like, ‘I felt really uncomfortable talking to the counsellor’ or ‘The therapist really weirded me out’. So we can use these statements like that (or in modified form) as items for our measure, and it might also trigger some ideas about further items, like ‘I felt really comfortable talking to the counsellor’ (a reverse of the first statement here), or ‘The therapist seemed really weird’ (a modification of the second statement).

  • Interviews and focus groups. Talk to people in the target population to see what terms they use to talk about the phenomena. For instance, an interview with young clients about their experiences of counselling (approved, of course, through the appropriate ethical procedures) might be an ideal way of finding out how they experience ‘awkwardness’ in therapy. What sort of words do they use to talk about it? How does it feel to them?

  • Dictionaries and thesauruses. Always a valuable means of finding different synonyms and antonyms for a phenomena.

Remember, what you are trying to do is to generate a range of items which are, potentially, a means of ‘tapping into’ your latent variable. Have a mixture of phrasings, with some items that are as closely worded to your latent variable as possible (for instance, ‘I felt awkward in therapy’), but other that might come at it from a different angle, providing ‘triangulation’ (for instance, ‘The interaction with my therapist seemed unusual’). It’s also good to try reversing some items (so, for instance, having items that are about not feeling awkward, as well as feeling awkward)—though having such items in a final scale is no longer considered essential.

At this point, you’ll also need to start thinking about your response categories: the ways that people score your items. For instance, do people rate the items on a 3- or 5-point scale, and what labels might you use to describe these different points? This is an enormous field of science in itself, and usually it’s best to keep it simple and use something that’s already out there so that it’s been tried and testing. For instance, if you decide to develop your own four point scale with labels like 1 = Not at all, 2 = A really small amount, 3 = Quite a bit, 4 = Moderately, 5 = Mostly, How do you know that Quite a bit means less to people than Moderately; and couldn’t the difference between 2 and 3 (A really small amount and Quite a bit) be a lot more than the difference between 4 and 5 (Moderately and Mostly)? So have a look at what other validated and well-used measures use as response categories and see if anything there suits. Two common ones are:

  1. = Strongly disagree

  2. = Moderately disagree

  3. = Mildly disagree

  4. = Mildly agree

  5. = Moderately agree

  6. = Strongly agree

    Or

  1. = Not at all

  2. = Only occasionally

  3. = Sometimes

  4. = Often

  5. = Most or all of the time

At this point, you’ll also need some idea of how you phrase the introduction to your measure. Generally, you’ll want to keep it as short as possible, but there may be some essential instructions to give, such as who or what to rate. For instance, for our racial microaggressions measures, we might want to say something like:

Please think of your relationship with your current therapist. To what extent did you experience each of the following?

In this instance, we might also consider it essential to say whether or not the clients’ therapists will see their scores, as this may make a big difference to their responses.

Testing items

Expert Review

The next stage of the measure development process is to pilot test our items. What we would do is to show each of our items to experts in the field (ideally experts by experience, as well as mental health professionals)—say between about 3 and 10 of them—and ask them to rate each of our potential items for how ‘good’ they are. We could do this as a survey questionnaire, on hard copy, or through questionnaire software such as Qualtrics. An example of a standardised set of questions for asking this comes from DeVellis’s brilliant book on scale development. Here, experts can be asked to rate each item on a four-point scale (1 = not at all, 2 = a little, 3 = moderately, and 4 = very well) with respect to three criteria:

  1. How well the item matches the definition of our latent variable (which the experts are provided with)

  2. How well formulated the item is for participants to fill in

  3. How well, overall, the item is suited to the measure

Once the responses are in, those items with the lowest ratings (for instance, with an average < 3) can be discarded, leaving only the most well formulated and suitable items to go forward for further testing and analysis.

Three Step Interviewing Technique

Something else that I’ve learnt, from Joel Vos, that can be really useful for selecting items in these early stages is called The Three-Step Test Interview. This essentially involves asking a few respondents (ideally the kind of people the measure is for) to ‘think aloud’ while completing the measure, and then to answer some interview questions about their experiences and perceptions of completing the measure. This, then, gives us a vivid sense of what the experience of completing the measure is like, and what’s working and what’s not. Through this process, for instance, it might become evident that certain items—even if the experts thought they were OK—don’t make much sense to participants, or are experienced as boring or repetitive. And respondents might also have ideas for how items can be better worded. Again, questions that don’t work well can be removed at this stage and, potentially, new or modified items could be added (though bear in mind they haven’t been through the expert review process).

Exploratory psychometrics

You’re now at the stage of sending your measure out to survey. The number of respondents you need at this stage is another field that is a science in itself. However, standard guidance is a minimum of 10 respondents per item, with other guidance suggesting at least 50 respondents overall if the aim is detect one dimension/scale, and 100 for two (see, for instance, here).

At this point, you almost certainly want to be inviting respondents to complete the measure online: for instance, through Qualtrics or Survey Monkey. Hard copies are an option, but add considerably to the processing burden and, these days, may make prospective participants less likely to respond.

Ideally, you want respondents to be reflective of the people who are actually going to use the measure. For instance, if it’s a measure intended for use with a clinical population, it’s not great if it’s been developed only with undergraduate students or with just your social media contacts. Obviously, too,it’s also important to aim for representativeness across ethnicity/race, gender, age, and other characteristics.

If you’ve got funding, one very good option here can be to use a Mechanical Turk service, such as Prolific. This is, essentially, a site where people get paid to complete questionnaires; and because it’s such a large pool of people, from all over the world, it means you’ve got more chance of recruiting the participants you need. We used this, for instance, to gather data on the reliability and validity of the Cooper-Norcross Inventory of Preferences (see write-up here), and it allowed us to get US and UK samples that were relatively representative in terms of ethnicity, gender, and age—not something we could have easily achieved just by reaching out to our contacts.

Once you’ve got your responses back, you’re on to the statistical analysis. The aim, at this point, is to get to a series of items that can reliably assess one or more latent dimensions, in a way that is as parsimonious as possible (i.e., with the fewest items necessary). This scale shortening process can be done in numerous ways, but one of the most common starting points is to use exploratory factor analysis (EFA).

EFA is a system for identifying the dimension(s) that underlie scores from a series of items. It’s a bit like putting an unknown liquid on a dish and then boiling it off to see what’s left: perhaps there’s crystals of salt, or maybe residues of copper or gold. EFA has to be done using statistical software, like SPSS or R (not Excel), and you need to know what you’re doing and looking for. On a 1-10 scale of difficult stats, it’s probably about a 5: not impossible to pick up but also does require some fair degree of training, particularly if you don’t have a psychology degree. What follows (as with all the stats below) is just a basic overview to give you an idea of the steps that are needed.

The first thing you do in EFA is to see how many dimensions actually underlie your data. For instance, the data from our ‘experiences of racial microaggression’ items may suggest that they are all underpinned by just one dimension: How much or how little people have experienced microaggressions from their therapists. But, alternatively, we may find that there were more latent dimensions underlying our data: for instance, perhaps people varied in how much they experienced microaggressions, but also the degree to which they felt hurt by the microaggressions they experienced. So while some people could have experienced a lot of microaggressions and a lot of hurt, others might have experienced a lot of microaggressions but not much hurt; and any combination across these two variables might be possible.

What EFA also does is to help you see how well different items ‘load’ onto the different dimensions: that is, whether scores on the items correlate well with the latent dimension(s) identified, or whether they are actually independent of all the underpinning dimensions on the measure. That way, it becomes possible to select just those items that reflect the latent dimension well, discarding those that are uncorrelated with what you have actually identified as a latent scale. At this point, it’s also common to discard those items that load onto multiple scales: what you’re wanting is items that are specifically and uniquely tied to particular latent variables. At this point, there’s many other decision rules that can also get used for selecting items. For instance, you might want items that have a good range (i.e., going the full length of the scale), rather than all scores clustering in the higher or lower regions; and the items also need to be meaningful when grouped together. So this process of scale shortening is not just a manualised one, following clearly-defined rules, but a complex, nuanced, and selective art: as much alchemy as it is science.

By the end of this exploratory process, you should have a preliminary set of items for each scale or subscale. And what you’ll then need to do is to look at the items for each scale or subscale and think about what they’re assessing: how will you label this dimension? It may be that the alchemical process leads you back to what you set out to find: a ‘prevalence of racial microaggressions’ dimension, for instance. But perhaps what crystallised out was a range of factors that you hadn’t anticipated. When we conducted our first Cooper-Norcross Inventory of Preferences study, for instance (see here), we didn’t really know what preference dimensions would emerge from it. I thought, for instance, that we might find a ‘therapist directed vs client directed’ dimension, as we did, but I was surprised to see that there was also a ‘focused challenge vs warm support’ dimension emerging as well—I had just assumed that therapist directiveness and challenge were the same thing.

Testing the Measure

As with exploratory measure development, there are numerous methods for testing the psychometric properties of a measure, and procedures for developing and testing measures are often iterative and overlap. For instance, as part of finalising items for a subscale, a researcher may assess the subscale’s internal reliability (see below) and, if problematic, adjust its items. These tests may also be conducted on the same sample that was used for the EFA, or else a new sample of data may be collected with which to assess the measure’s psychometric properties.

Two basic sets of tests exist that most researchers will use at some point in measure development research: the first concerned with the reliability of the measure and the second concerned with its validity.

Basic Reliability Tests

The reliability of a measure is the extent to which it produces consistent, reproducible estimates of an underlying variable. A thermometer, for instance, that gave varied readings from one moment to the next wouldn’t be much use.

  • Internal consistency is probably the most important, and frequently reported, indicator of a scale’s ‘goodness’ (aside from when the measure is idiographic). It refers to the extent that the different items in the scale are all correlating together to measure the same thing. If the internal reliability is low, it means that the items, in fact, are not particularly well associated; if high, it means that they are all aligned. Traditionally, internal consistency was assessed with a statistic called ‘Cronbach’s alpha (α)’, with a score of .7 or higher generally considered adequate. Today, there is increasing use of a statistic called ‘McDonald’s omega (ω)’, which is seen as giving a less biased assessment.

  • Test-retest reliability is very commonly used in field of psychology, but is, perhaps, a little less prevalent in the field of counselling and psychotherapy research, where stability over time is not necessarily assumed or desired. Test-retest reliability refers to the stability of scores over a period of time, where you would expect people to score roughly the same on a measure (particularly if it is a relatively stable trait). If respondents, for instance, had wildly fluctuating scores on a measure of self-esteem from one week to the next, it would suggest that the measure may not be tapping into this underlying characteristic. Test-retest stability is often calculated by simply looking at the correlation of scores from Time 1 to Time 2 (an interval of about two weeks is typically used), though there are more sophisticated statistics for this calculation. Assessing test-retest reliability requires additional data to be collected after the original survey—often with a subset of the original respondents.

  • Inter-rater reliability is used where you have an observer-completed measure. Essentially, if the measure is reliable, then different raters should be giving approximately the same ratings on the scales. In our assessment of an auditing measure for person-centred practice in young people, for instance (see here), we found quite low correlations between how the raters were assessing segments of person-centred practice. That was a problem, because if one rater, on the measure, is saying that the practice is adherent to person-centred competencies, and another is saying it isn’t, then it suggests that the measure isn’t a reliable means of assessing what is and is not a person-centred way of working.

Basic Validity Tests

The validity of a measure is the extent to which it measures the actual thing that it is intended to. Validity can be seen as the ‘outward-facing’ element of a measure (how it relates to what is really going on in the world), whereas reliability can be seen as the ‘inward-facing’ element (how the different parts within it relate together).

  • Convergent validity tends to be the most widely emphasised, and reported, test of validity in the counselling and psychotherapy research field. It refers to the extent that scores on the measure correlate with scores on a well-established measure of a similar construct. Suppose we were developing a measure to assess how prized clients feel by their therapists. No measures of this exact construct exist out there in the field (indeed, if it did, we wouldn’t be doing this work), but there’s almost certainly other scales, subscales, or even individual items out there that we’d expect our measure to correlate with: for instance, the Barrett-Lennard Relationship Inventory’s ‘Level of Regard’ subscale. So we would expect to find relatively high correlations between scores on our new prizing measure and those on the Level of Regard subscale, say around .50 or so. If the correlations were zero, it might suggest that we weren’t really measuring what we thought we were. But bear in mind that correlations can also be too high. For instance, if we found that scores on our prizing measure correlated extremely closely with scores on Level of Regard (> .80 or so), it would suggest that our new measure is pretty redundant: the latent variable we were hoping to tap has already been identified as Level of Regard. Assessing convergent validity means that, in our original survey, we might also want to ask respondents to complete some related measures. That way, we don’t have to do a further round of surveying to be able to assess this psychometric property.

  • Divergent validity is the opposite of convergent validity, and is essentially the degree to which our scale or subscale doesn’t correlate with a dimension that should be unrelated. For instance, our measure of how prized clients feel wouldn’t be expected to correlate against a measure of their degrees of extraversion, or levels of mental wellbeing. If they did, it would suggest that our measure is measuring something other than we think it is. Measures of ‘social desirability’ are good tools to assess divergent validity against because we really don’t want our measure to be associated with how positively people try to present themselves. As with assessing convergent validity, assessing divergent validity means that we may need to add a few more measures to our original survey, if we don’t want to go through a subsequent stage of additional data collection.

  • Structural validity is the degree to which the scores on the measure are an adequate reflection of the dimensions being assessed. EFA, as discussed above, can be used to identify one or more underlying dimensions, but this structure needs validating in further samples. So this means collecting more data (or splitting the original data into ‘exploratory’ and ‘confirmatory’ subsamples), and then the new data can be analysed using a procedure called confirmatory factor analysis (CFA). CFA is a complex statistical process (about a 9 on the 1-10 scale), but it essentially involves testing whether the new data fits to our ‘model’ of the measure (i.e., its hypothesised latent dimension(s) and associated items). CFA is a highly rigorous check of a measure, and it’s a procedure that’s pretty much essential now if you want to publish a measure development study in one of the higher impact journals.

  • Sensitivity to intervention effects is specific to outcome measures, and refers to the question of whether or not the measure picks up on changes brought about by therapy. We know that therapy, overall, has positive benefits, so if scores on a measure do not show any change from beginning to end of intervention, it suggests that the measure is not a particularly valid indicator of mental wellbeing or distress. To assess this sensitivity, we need to use the measure at two time points with clients in therapy: ideally at the start (baseline) and at the end (endpoint). Measures that show more change may be particularly useful for assessing therapeutic effects. For instance, in our psychometric analysis of a goal-setting measure for young people (the Goal Based Outcome Tool), we found that this measure indicated around 80% of the young people had improved in therapy, as compared with 30% for the YP-CORE measure of psychological distress.

Advanced Testing

…And there’s more. That’s just some of the basic psychometric tests and, like I said earlier, there seems to be new ones to catch up with everyday, with numerous journals and books on the topic. For instance, testing for ‘measurement invariance’ seems to becoming increasingly dominant in the field, which uses complex statistical processes to look at whether the psychometrics of the measures are consistent across different groups, times, and contexts (this is about a 15 out of 10 for me!). And then there’s ‘Rasch analysis’ (see here), which uses another set of complex statistical procedures to explore the ways that respondents are scoring items (for instance, is the gap between a score of ‘1’ and ‘2’ on a 1-5 scale the same as the gap between ‘3’ and ‘4’?). So if you’re wanting to publish a measure development study in the highest impact journals, you’ll almost certainly need to have a statistician—if not a psychometrician—on board with you, if you’re not one already.

Developing benchmarks

Once you’ve got a reliable and valid measure, you may want to think about developing ‘benchmarks’ or ‘cutpoints’, so that people know how to interpret the scores from it. This can be particularly important when you’re developing a clinical outcome measure. Letting a client know, for instance, that they’ve got a score of ‘16’ on the PHQ-9 measure of depression, in itself, doesn’t tell them too much; letting them know that this is in the range of ‘moderately severe depression’ means a lot more.

There’s no one way of defining or making benchmarks. For mental health outcome measures, however, what’s often established is a clinical cut-off point (which distinguishes between those who can be defined as being in a ‘clinical range’ and those in a ‘non-clinical range’); and a measure of reliable change, which indicates how much someone has to change on a measure for it to be unlikely that this is just due to chance variations. For instance, on the Young Person’s CORE measure of psychological distress, where scores can vary from 0 to 40, we established a clinical cut-off point of 10.3 for males in the 11-13 age range, and a reliable change index of 8.3 points (see here). The calculations for these benchmark statistics are relatively complex, but there are some online sites which can help, such as here. You can also set benchmarks very simply: for instance, for our Cooper-Norcross Inventory of Preferences, we used scores for the top 25% and bottom 25% on each dimension as the basis for establishing cut-off points for ‘strong preferences’ in both ways.

the Public domain

Once it’s all finalised and you’re happy with your measure, you still need to think about how you’re going to let others know about it. There’s some journals that specifically focus on the development of measures, like Assessment, though they’re by no means easy to get published in. Most counselling and psychotherapy journals, though, will publish measure development studies in the therapy field, and that puts your measure out into the wider public domain.

At this stage you’ll also need to finalise a name for your measure—and also an acronym. In my experience, the latter often ends up being the toughest part of the measure development process, though sites like Acronymify can help you work out what the options might be. Generally, you want a title that is clear and specific to what your measure is trying to do; and a catchy, easy-to-pronounce acronym. If the acronym actually means or sounds something like what the measure is about—like ‘CORE’—that’s even better.

If there’s any complexities or caveats to the measure at all in terms of its use in research or clinical practice, it’s good to produce really clear guidelines for those who want to use it. Even a page or so can be helpful and minimise any ambiguities or potential problems with its application. Here is an an example of the instructions we produced for our Goals Form.

It can also be great to develop a website where people can access the measure, its instructions, and any translations. You can see an example of this for our C-NIP website here.

Regarding translations, its important that people who may want to translate your measure follow a standardised procedure, so that it stays as consistent as possible with the original measure. For instance, a standard process is to ‘back translate’ an initial draft translation of the measure to check that the items are still meaning the same thing.

In terms of copyright, you can look at charging for use of the measure, but personally I think it’s great if people can make these freely available for non-commercial use. But to protect the measure from people amending it (and you really don’t want people doing their own modifications of your measure) you can use one of the Creative Commons licenses. With the measures I’ve been involved with, we’ve used ‘© licensed under the Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International (CC BY-NC-ND 4.0)’ so that others can use it freely, but can’t change it or make money from its use (for instance, by putting it on their own website and then charging people to use it).

Conclusion

At the most advanced levels, measure development and testing studies can be bewildering. Indeed, even at the most basic level they can be bewildering—particularly for those who are unfamiliar with statistics. But don’t let that put you off. There’s a lot of the basic item generation and testing that you can do without knowing complex stats, and if you’re based at an institution there’s generally someone you can ask to help you with the harder stuff. There’s also loads of information that you can google. And what you get at the end of it is a way of operationalising something that may be of real importance to you: creating a tool which others can use to develop knowledge in this field. So although measure development research can feel hard, and like a glacially slow process at times, you’re creating something that can really help build up understandings in a particular area—and with that the potential to develop methods and interventions that can make a real difference to people’s lives.

Acknowledgements

Photo by Tran Mau Tri Tam ✪ on UnsplashDisclaimer

 The information, materials, opinions or other content (collectively Content) contained in this blog have been prepared for general information purposes. Whilst I’ve endeavoured to ensure the Content is current and accurate, the Content in this blog is not intended to constitute professional advice and should not be relied on or treated as a substitute for specific advice relevant to particular circumstances. That means that I am not responsible for, nor will be liable for any losses incurred as a result of anyone relying on the Content contained in this blog, on this website, or any external internet sites referenced in or linked in this blog.

Getting Published in Higher Impact Journals: Some Pointers

Let me start with some caveats. First, I couldn’t claim to be an expert in getting published in high impact journal. I’ve had some successes—with articles, for instance, in journals like Lancet Child & Adolescent Health, Journal of Consulting and Clinical Psychology, and Clinical Psychology: Science & Practicebut also numerous failures, including a really disappointing rejection just a few nights ago. Compared to the Michael Barkhams or Clara Hills of this world, I’m a mere novice.

Second, not everyone should, or does, want to get published in higher impact journals. Indeed, for many people, it’s an elitist, Global North-centred system that excludes non-academics and those who aren’t willing to comply with a positivist, scientistic mindset. So this blog is not suggesting that higher impact journals are good to get published in, or better than other journals, or something that all academics (and non-academics) should be aspiring to. But it is written on the basis that, for some academics, getting published in higher impact journals is important: for their careers and, perhaps, more importantly, for the maintenance and development of counselling and psychotherapy programmes in higher education (HE). In the 2000s we witnessed many counselling courses at HE institutes get closed down and, in some cases, this was because the teams were seen as not producing enough research output at a sufficiently high level. So for the maintenance and enhancement of counselling in the UK and globally, it may be really important for academics to be publishing at the highest possible level: not just for them but for the counselling community as a whole.

What is an ‘impact factor’?

So what do I mean by higher impact journals? Well, for those who don’t know, impact is essentially an indicator of the status of a journal, and it’s operationalised in terms of the amount of citations that the average article will have over the two year period from publication (for a more general guide to publishing in therapy journals, see here). You can go to any journal home page and if you look under tabs like ‘Journal Metrics’ you’ll find the impact factor (sometimes specifically called the ‘2-year impact factor’): for instance, for Psychotherapy Research it’s 3.768. This means that, on average, an article in Psychotherapy Research has been cited 3.768 times by other articles, in any other academic journal, over that two year period. Higher is, of course, ‘better’, meaning that articles in that journal are being more widely drawn on by other members of the academic community.

Impact factors for journals vary a lot by particular disciplines. For instance, in the medical and scientific fields, there’s quite a quick turn-over of articles: they come out quickly and then are rapidly drawn on by other members of those disciplinary communities. That means that journals like Science Robotics or Cancer Research can easily have impact factors of 10 or more. In the counselling and psychotherapy research field, impact factors tend to be a bit more modest, though they have increased in recent years. They range from about 1 (e.g., British Journal of Guidance and Counselling) to 5 or more (e.g., Journal of Consulting and Clinical Psychology), with some of the more psychiatric journals even higher (e.g., Lancet Psychiatry with an impact factor of 27.083).

So by ‘higher impact journals’, I mean counselling and psychotherapy journals with an impact factor of, say, about 2.5 or more. In many instances, these are US-based journals; and, as above, in most cases they are of a relatively positivist, scientistic mindset. There are journals that are more experientially- or constructionistically-focused (like the European Journal of Psychotherapy & Counselling and BACP’s Counselling & Psychotherapy Research), but mostly they either have a low impact factor or none at all.

Just to note, not having a formal impact factor doesn’t mean that there are no citations to papers in that journal. To have a formal impact factor, a journal needs to be recognised by an organisation called Clarivate (formerly Thomson-Reuters), and they are very selective about the journals that are recognised. Applying for recognition can also be a very slow process. So there are some very good journals, like the Journal of Psychotherapy Integration and Counselling Psychology Quarterly, that don’t have a formal impact factor. However, these days, such journal may calculate their own 2-year impact factor and present it on their site; so a single organisation’s monopoly over impact factors seems, thankfully, to be waning.

‘Playing the game’

As indicated above, nearly all the higher impact journals in counselling and psychotherapy can be quite positivist, scientistic, and realist in their mindset. And the reality is, that’s not likely to change (at least, not in the short term). They’re bombarded with articles and can pick and choose what they want to publish, often with rejection rates of 70% or more. So if you approach them, say, with an autoethnographic study of authenticity in therapy, you can argue with them until you are blue in the face about the importance of reflexivity and the social construction of reality, but they are unlikely to budge. They’re generally quite conservative: they have their ways of doing things, and they simply don’t need to change—whether or not they should. So, as a first and overarching pointer, if you want to publish in these journals, you generally need to ‘play by their rules’. It’s an uncomfortable reality for many of us, but it’s the way things are.

Learn Stats, or Find a Statistician

And this is a first implication of playing by their rules: if you want to stand a good chance of getting published, having a high quality statistical analysis is often a good way in. Most of the higher impact journals prefer quantitative articles to qualitative; indeed, some have explicitly said that they’re not interested in publishing qualitative articles (primarily because they’re seen as lacking generalisability because of low sample sizes). And the stats we’re talking about here is more than just some means and standard deviations (see blog on quantitative analysis, here). We’re talking structural equation modelling, multilevel analysis, cross lag panel designs… the kind of stats that, I know for myself, I can only just about understand—let alone do.

So options are to learn, in depth and detail, one particular statistical method (or a few) and then apply that. Or, and this is what most of us do, bring on board a statistician who is able to do analyses at the requisite level. That latter strategy is fairly pervasive across the research field: you collaborate with someone who specialises in statistics, and pass on the data to them for an in-depth analysis. If you’re at a university, there may well be someone in your department or faculty that has that role—or you can try linking with stats experts at other universities. Generally, someone with an in-depth understanding and specialism in stats is always going to do better than a non-stats person trying to learn a new, specialised method—unless that person really love maths and stats, and has the time and inclination to learn complex methodologies.

Control, Control, Control

Controlled experimental studies (where, for instance, some participants are allocated to an intervention and some are not) are the lifeblood of the psychological field, and they’re very popular amongst the higher impact therapy journals too. Why? Because they are seen as the ‘gold standard’ means of demonstrating causal effects. Everything else—pre-post studies, qualitative research, observational studies, etc.—tend to be seen as correlational only. So if there’s a way of conducting a controlled study in your area of interest (and doing it in a highly rigorous way using, for instance, CONSORT [Consolidated Standards of Reporting Trials] guidelines), there’s a good chance of getting it published. Ideally, you’ve also got the numbers of participants to make the trial ‘powered’ at an adequate level (see more about statistical ‘power’, here). That could be 100 or so participants per condition; but even if you cannot achieve that, a ‘pilot’ or ‘feasibility’ study should be possible and publishable (albeit in lower impact journals). Also, a controlled study doesn’t have to be a trial of a full intervention. For instance, you might test the use of a particular homework exercise, or of using visualisations as part of the therapeutic work. Or perhaps you’d give some clients full information about the therapy, and others only basic information, to see if that makes a difference. Whatever you’re line of research interest, controlled trials are nearly always possible; and you can normally also conduct qualitative research alongside them to get a more experiential, in-depth understanding of the intervention effects.

If you’re doing qual, do it systematically

Having said all that, there are higher impact journals who will consider—and do publish—qualitative research, so it’s by no means impossible to get qualitative research published. But, even so, some journals can tend to assess it with a fairly ‘quantitative’, positivistic mindset (what’s been called ‘small q research’). That means that studies with larger samples, and some quantification of themes, may be preferred (for instance, ‘framework analysis’). Also, very importantly, evidence of inter-coder reliability—or some triangulation of coding—may be expected. That means having some means of demonstrating that the coding process wasn’t dependent on just the ratings of one person, but has some ‘objective’ reality. That could be done, for instance, by having two or more analysts working independently on a set of transcripts, and then agreeing together the coding. Or having a group conduct the analysis, as in consensual qualitative research. Another way of doing this would be to have a second coder do some coding and then compare this with the main coder, producing some statistic for inter-rater reliability (such as Cohen’s kappa).

There are exceptions to this. For instance, some higher impact psychotherapy research journals have published interpretative phenomenological analysis studies where there is limited or no evidence of inter-coder reliability (see an excellent example here; an IPA study on the therapeutic relationship in CBT for adolescents). And, hopefully, such kinds of studies are becoming more common in the higher impact journals. However you do it though, what’s essential is that the qualitative research is conducted systematically. That can mean sticking closely to an established, defined methodology (such as reflexive thematic analysis); or, if you are using a mixed or new method, explaining the rationale and the procedures in that very clear. It’s also essential that the methodology is in line with the aims of your research: Why use this method to be trying to answer these research questions? And finally, do make sure you really process the data: spend time with it, examine it in depth, work out what it is really telling you. There really is no easy way to sophisticated knowledge.

Systematic Reviews

If you’re not keen on stats, another option that the higher impact journals are often open to is systematic reviews. These tend to be popular because they are generally well-cited, boosting the impact factor of the journal itself (remember, journal editors will be thinking about how often your article is likely to be cited—they want to keep their impact factors up!). Again, journals often prefer statistically-based reviews of the literature (i.e., meta-analyses), but they can also be open to narrative and other types of in-depth review. If you want to increase your chances of being accepted, though, again it needs to rigorously follow well-established methods for conducting such research. So, for instance, base the review around the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) guidelines: with, for instance, multiple coders during the study selection process, flow charts, and formal assessment of bias procedures.

Plan from the Start

One of the biggest mistakes I’ve made in my academic career is starting research projects thinking, ‘Oh, I’m sure this will generate some interesting data,’ and then getting to the end of the data collection process and realising that, actually, it is not quite the right data for a high quality publication. So I’ve learnt that it is really important, if you want to publish in a higher impact journal, to plan right from the start: What the specific aims of the study are, What you want to contribute to the literature, and Where you want to publish it. Different journals have different interests, so having a specific journal in mind for publication (and knowing the kinds of articles that journal publishes) helps to ensure that you are progressing along the required tracks. Michael Barkham, Professor of Clinical Psychology at the University of Sheffield and a world-leading psychotherapy researcher, says:

I always tell folk when working out plans and methods to imagine how they are going to write this up for a high-quality journal—so your ‘headset’ is the final product that then helps to shape the process. The point here is that having a ‘headset’ of a high-quality journal output from the start becomes the guiding process for delivering better quality research (although it doesn't guarantee publication success).

If you can set your plans out in a study protocol or study analysis plan and then ‘pre-register’ it on a site like the Centre for Open Science, it can then also really help to convey to journal reviewers the rigour with which your study has been conducted. And consult: send your protocol to people who have published in those journals and see what they think. Better to make tweaks at this earlier stage then get all the way to the end of data collection before you realise your method has some severe limitations.

It Takes time

My highest impact journal publication (that I’ve led on) took about six years from conception to publication (the ETHOS study: an RCT on the effectiveness of school counselling), and there was about a decade before that of previous work in the school counselling field. To get published in these journals, you really need to be at the absolute forefront of a field—to know everything there is to know about it—and then to conduct research that is going to significantly take that field forward. So it’s not something you can do overnight, it takes times: to build up expertise, to set out a research study, to gather the requisite amount of data, to write and to finesse a paper with multiple drafts and re-drafts. Generally, developing expertise in a specialist field—and then publishing and publishing on that—is a better strategy that trying to be a generalist (and I say that from my own experience of trying to cover too much). If you get spread too thin, there’s no way that you can be at the forefront of every field. Rather, choose a field—like counselling for people with autism, or empathy in the therapeutic relationship, or moments of deep connection in therapy—and work at it and work at it and keep researching, reading, and linking up with other leading people in that field.

Work in a Team

And that links to team work. It’s very rare these days that people publish papers in leading journals that they’ve written alone. Rather, there’s often a list of three or more—and, in some of the scientific papers, hundreds—of co-authors. That’s because people have developed the research as a team, and having multiple people working with you—within the same institution, or across institutions—is often essential in bringing together the expertise needed to publish research at the highest level. If you’ve got a team, for instance, you can have world-leading expertise in research design, and in statistics (as above), and then in a particular intervention—how many people have all that in themselves? In a team, everyone can help you, as a whole, to ensure you’re at the forefront of that field. And the great thing is, for academic auditing systems like the Research Excellence Framework (REF), it doesn’t matter whether you are the first author or just one of the co-authors, it all counts towards your published ‘outputs’.

If you’re yet to publish in higher impact journals, joining up with (and contributing to) a team of more experienced researchers can give you a crucial toe-hold in this world, says Michael Barkham. ‘It gives you a connection from which you can learn how this work can be done—start as a small cog and progress from there.’ So this is about being a junior partner with more senior colleagues: for instance, conducting the qualitative analysis in a primarily quantitative trial, or being part of a coding team. That way, you can learn the craft of higher impact journal publication, and take things forward from there.

Mentoring

Allowing oneself to be mentored by a well-published researcher, suggests Michael Barkham, can be an essential part of this process. This might be an informal arrangement between colleagues (within or across universities); or a more formal arrangement, such as a PhD programme. In the US and in much of Europe, this is exactly the system that produces so many well-published young researchers. Senior academics take on a small handful of PhD students, and work closely with them—over several years—to produce high impact research. By the time the students have left the PhD programme, they have learnt the skills and requirements of higher impact journal production, and are ready to ‘fly the nest’ and lead research on their own.

Funding

Finding the time to develop that expertise isn’t always easy: not for you, and often not for your potential colleagues. So if you can get funding to pay for your time, and to bring in additional researchers (for instance, through grants from the Economic and Social Research Council), that’s nearly always a great basis for developing high quality research programmes. Getting that funding is tough, no doubt, but there’s numerous potential sources (see, for instance, Research Professional); and the more you can develop expertise in a particular area, the more successful you’re likely to be. Again, it’s about ‘playing the long game’: not one-off attempts at high impact publication, but a long and sustained development of expertise in a particular area, that will eventually bear high quality fruits.

Show Added Value

For the higher impact journals, doing a well-conducted study is not enough. Great, so you did an IPA study the right way, or conducted some high quality statistical analysis, but what does it all mean? Journals are looking for papers that really take the field forward, so you have to make it explicit in your paper what it is that you’re adding to what was already known. Maybe you’ve discovered that clients really value a particular form of therapist self-disclosure, or that relational depth is a key predictor of therapeutic outcomes. Of course, you need to be honest about the limitations of your research; but if it’s all limitations, and null results, and ambiguities, higher impact journals may be more likely to send it back. Why publish something that, at the end of the day, doesn’t tell us too much when they can publish papers that will have clear and robust implications for practice, training, or research?

Don’t Give Up

Maybe this is all a bit bleak, but it’s written from (bitter) experience. And, having said all that, it is really possible to get published in some of the best journals in the world—you just need to be smart and strategic about it. And, perhaps more than anything, you need to be resilient and keep going in the face of rejection (…after rejection, after rejection, after rejection—says a man who is still smarting after his rejection email a few days ago). These higher impact journals can be brutal in assessing work. And they can be infuriating in imposing standards that you may think are totally wrong. But if you keep at it, and learn from the feedback you’re receiving, there’s a good chance that sooner or later you’ll succeed.


Publishing in higher impact journals is hard, it’s really hard. So it can’t be an afterthought or something you think you’ll have a go at as a corollary to something else. For instance, if you’re doing some research primarily as ‘personal development’, but then think, ‘Well, let’s see if the top journals are interested in what I’ve produced,’ the chances are, they won’t be. Rather, as above, if you want to publish in these higher impact journals, this has to be your focus and your goal, your ‘headset’, from the start. And there is a point here, perhaps, for the wider UK counselling and psychotherapy community. If we want to be part of this higher impact publishing world, then organisations (like BACP and UKCP), and counselling academics, need to be oriented this way from the off. We need, for instance, BACP to work with counselling academic groups—as I know they are trying to do—to set up specialised research programmes, with mentoring and PhDs so that there is a sustainable programme of research at the highest possible level.

Should we be compromising to get published in these higher impact journals? I read a brilliant paper by Virginia Braun and Victoria Clarke, developers of thematic analysis, on the London Underground into work this morning, and it made me feel like, ‘No, what the hell, we should be developing qualitative and phenomenological and reflexive inquiry in a way that we believe in, and we can and should be getting that work published—at the highest possible level.’ Perhaps so. I honestly don’t know. Maybe the mountain will come to us. I guess, in my own career and in the career of those in the UK counselling field, I just haven’t seen that happen too much. I was a radical in my youth. Now, as I get older and older, I see more and more the virtues of compromise. Or, perhaps more positively, compromise as a means of achieving, ultimately, more radical results. And also, compromise as a respect to those who have different views and takes on the world. There can be an arrogance in radicalism—there was in my own radicalism—which contradicts its very essence.

Finally: writing this blog makes me realise that, ultimately, getting published in higher impact journals is a process, not an outcome: publishing in these top journals needs to be embedded in a wider programme of research, development, and impact. You can’t just go off, on your own, and write a high quality paper (or, at least, I can’t). Rather, it’s about specialising, focusing, developing expertise in a particular field—and being an integral part of a community that is doing the same thing. Then, it’s not that you want to publish in the higher impact journals, it’s that they want to publish you.

Acknowledgements

Thanks to Michael Barkham for input and guidance. Photo by alex starnes on Unsplash

Disclaimer

 The information, materials, opinions or other content (collectively Content) contained in this blog have been prepared for general information purposes. Whilst I’ve endeavoured to ensure the Content is current and accurate, the Content in this blog is not intended to constitute professional advice and should not be relied on or treated as a substitute for specific advice relevant to particular circumstances. That means that I am not responsible for, nor will be liable for any losses incurred as a result of anyone relying on the Content contained in this blog, on this website, or any external internet sites referenced in or linked in this blog.

Quantitative Analysis: Some Pointers

When it comes to counsellors and psychotherapists, everyone hates stats. Well, almost everyone. Aside from a few geeks like myself who would prefer nothing better than sitting in front of an Excel spreadsheet for days.

…Oh yes: and, then, there’s also the funders, commissioners, and policy-makers who all rely almost exclusively on the statistical analysis of data. And that creates a real tension. Most of us don’t come into therapy to do statistical analysis. We want to engage with people—real people—and studying people and processes by numbers can feel like the most de-humanising, over-generalising kind of reductionism. But, on the other hand, if we want to have an impact on the field and influence policy and practice, then we do need to engage with quantitative, statistical analysis. Or, at least, understand what it is saying and showing. If not, there’s a danger that those therapies that are most humanistic and anti-reductionistic are also those that are most likely to get side-lined in the world of psychological therapy delivery.

And there is also another, less polarised, way of looking at this. From a pluralistic standpoint, no research method—like no approach to therapy—is either wholly ‘right’ or ‘wrong’ (see our recent publication on pluralistic research here). Rather, different methods of research and analysis are helping in asking different questions at different points in time. So if you are asking, for instance, about the average cost of a therapeutic intervention; or whether, on average, a client is more likely to find Therapy A or Therapy B more helpful; then it does make sense to use statistics. (But if you wanted to know, for instance, how different clients experienced Therapy A, then you’d be much better off using qualitative methods).

This blog presents a very basic introduction to terms and concepts in quantitative analysis. This may be helpful if you are wanting to present some basic statistical analyses in a research paper, or if you are reading quantitative research papers and want to get more of a grasp on what they are meaning and doing. You can find many books and guides on the internet that give more in-depth introductions to quantitative analysis, one of the most popular being Andy Field’s Discovering Statistics Using IBM SPSS Statistics.

Quantitative analysis and statistical analysis are essentially the same thing (and will be used synonymously in this blog): the analysis of number-based data. The principle alternative to quantitative analysis is qualitative analysis, which refers to the analysis of language-based data.

Descriptive Statistics

There are two main sorts of quantitative analysis. The first is descriptive statistics. This is where numbers are used to show what a set of data looks like (as opposed to testing particular hypotheses, which we’ll come on to). Descriptive statistics may be used in a Results section to present the findings of a study but, even if you are doing a qualitative study, you may use some descriptive statistics to present some data about your participants. So always worth knowing.

Frequencies

Probably the most basic statistic is just saying how many of something there are: for instance, how many participants you had in a study, or how many of them were BAME/White/etc. There are two basic ways to do this:

  • Count. ‘There were nine participants in the study; three of them were of a black or minority ethnicity and six were white.’ Count is just the number of something, and about as simple as statistics gets.

  • Percentage. Percentage is the amount of something you would have if there was 100 in total. It’s a way of standardising counts so that we can compare them. For instance, if we had three BAME participants out of nine total participants in one study; and ten BAME participants out of 1000 total participants in a second study; the count of BAME participants in the second study is higher, but actually they were more representative in the former. We work out percentages by dividing the count we’re interested by the total count, then multiplying by 100. So our percentage in the first study is 3/9 * 100 (‘/’ means ‘divide by’, ‘*’ means ‘multiply by’), which is 33.3%; and in our second study is 10/1000 * 100, which is 1%. 33.3% vs. 1%: that really shows us a meaningful difference in representation across the two studies. Percentages are easy to work out by Excel: just do a formula where you divide the number in the group of interest by the total number, and multiple by 100. Only do percentages when it’s needed though: that is, when it would be hard for the reader to work out the proportion otherwise. With small samples (less than ten or so) you probably don’t need it. If we had, for instance, one White and one BAME participant, it’s a bit patronising to be told that there’s 50% of each!

Averages

One way of pulling together a large set of numerical data is through averages. This is a way of combining lots of bits of data to give some indication of what the data, overall, looks like. There are three main types of averages:

  • Mean. This is the one that you come across most frequently, and is generally the most accurate representation of the middle point in a set of data. The mean is the mathematical average, and is worked out by adding up all the scores in a set of data and then dividing by the number of data points. For instance, if you had three young people who scores on the YP-CORE measure of psychological distress (which ranges from 0 to 40, with higher scores meaning more distress) were 10, 15, and 18, then we could work out the mean by adding the scores together (which gives us 43) and then dividing by the number of scores (which is 3). So the mean is 43/3 = 14.3. Whenever we have several bits of data along the same scale—for instance ages of participants in a study, or scores of participants on a measure of the alliance—it can be useful to combine it together using the mean. Means are easy to do on Excel using the function AVERAGE. Note, don’t worry about lots and lots of decimal points. Really, for instance, that the mean above is 14.3333333333333333333333333 etc but no-one needs to know that level of detail. It just looks like we are trying to be clever and actually makes it harder for the reader to know what is going on! So normally one decimal point is enough (unless the number is typically less than 1.0, in which case you could give a couple of decimal points).

  • Median. Sometimes our data might have an usual distribution. Supposing, for instance, that we did a study and our participants ages were 20, 22, 23, 24, and 95. Well, the mean here would be 36.8 years old, but it doesn’t seem to describe our data very well because we have one ‘outlier’ (the 95 year old) who is very different from the other participants. So an alternative kind of average is the median, which is where we line up our values in a consecutive sequence, and then identify the middle. In this instance, we have five values and the middle one is 23 years old. The MEDIAN function on Excel is also very easy to use, and is a useful way of describing our data when there isn’t too much of it or it’s not smoothly spread out. If the mean and median of a set of values are very different, it’s normally helpful to give both—less important if they are virtually the same.

  • Mode. Let’s be frank, the mode is like the useless youngest sibling of the central tendency family: it doesn’t really tell you much and doesn’t get used very often. The mode is just the most common response. So, for instance, if we had YP-CORE scores of 20, 20, 23, 25, and 40 the mode would be 20 because there are two of those scores and one of every other one. Not much use, huh! But some times it can be quite informative. For instance, it’s an interesting fact that the modal number of sessions attended at many therapy services is 1. So even though the mode and median may be closer to 6 or so sessions, it’s interesting to note that the most common number of sessions attended is much less. MODE can be shown in Excel, but only report it if it adds something meaningful to what you are presenting.

Spread

Say you had a group of people who were aged 20, 30, and 40 years old. Then you had a second group that were aged 29, 30, and 31 years old. If we just gave the mean or the median of the groups, they’d actually be the same: 30 years old. But, clearly, the two sets of data are a bit different, because the first one is more spread out than the second one. So, if we want to understand a dataset as comprehensively as possible, with as few as possible figures, then we also need some indication of spread.

  • Range. The range is the simplest way of giving an indication of the spread of a dataset, and just means giving the highest and lowest values. So, for instance, with the first dataset above you might say: ‘Mean = 30 years old, range = 20-40 years old’. That can be pretty informative, though for larger datasets the highest and lowest numbers don’t tell us much about what is in the middle.

  • Standard deviation. The standard deviation, or SD, is an indication of the spread of a dataset. In contrast to the frequencies or central tendencies, it’s not a number that intuitively means much, but it’s essentially the average amount that the values in a dataset vary from the mean. So in the first group above, the standard deviation is 10 years and in the second group it’s 1 year. Essentially, a higher standard deviation means more spread. Pretty much always, if you’re giving a mean you’ll also want to give the standard deviation; so, in a paper, you’ll see something like: ‘Mean = 30 years old (SD = 10)’. Means look pretty naked without an SD. But it’s not easy to work out yourself, and you’ll need to use something like Excel that can calculate it using the function STDEV.

  • Standard error. This is getting a bit more complicated, and you’re unlikely to need standard error (SE) if you’re just presenting some simple descriptive statistics, but it is worth knowing about because it’s the basis for a lot of subsequent analyses. Let’s say we’re interested in the levels of psychological distress of young people coming in to school counselling, and we use the YP-CORE to measure it. We get an average level of 20.8 and a standard deviation of 6.4 (this is what we actually got in our ETHOS study of school-based counselling). So far so good. But, of course, this is just one sample of young people in counselling, and what we really want to know is what the average levels of distress of all young people coming into counselling is: the population mean (so sample is the group we are studying, and population is everyone as a whole). So how good is our mean of 20.8 at predicting what the population mean might actually be? OK, so here’s a question: if that mean came from a sample of 10,000 young people, or if it came from a sample of five young people, which would give the most accurate indicator of the population mean (all other things being equal)? Answer (I hope you got this), from the sample of 10,000 people. Why? Because in the sample of five young people, any individual idiosyncrasies could really influence the mean; whereas in a much larger sample these are likely to get ironed out. So the standard error is an indication of how much the sample mean is likely to vary from the true population mean, and it’s worked out by dividing the standard deviation by the sample size (square-rooted). Don’t worry about why it’s the square root (the number that, when multiplied by itself, gives that value—for instance the square root of nine is three). But it just means that the larger a sample, the smaller the standard error gets: indicating that it varies around the true population mean by a smaller amount. Phew!

  • Confidence intervals. Again, the standard error, as a statistic, isn’t a number that intuitively means much. One thing that is often done with it, however, is to work out the confidence intervals around a particular mean. The confidence interval is our guestimate of where the true population mean is likely to lie, given our sample mean. And it’s always at a particular level of confidence, normally 95% (or sometimes 99%). So if you see something like ‘Mean YP-CORE score = 20.8, 95%CI = 19.5 - 22.1)’ it’s telling us that we can be 95% certain that the true population mean for YP-CORE scores of young people coming into counselling is between 19.5 and 22.1. Pretty cool, and confidence intervals are used more and more these days, because there’s a move from pretending we know precisely what a population mean is to being more cautious in suggesting whereabouts it might lie. Confidence intervals aren’t too difficult to calculate—for 95% CIs, you add, and take away, 1.96 * the standard error—but, like standard errors, there’s no automatic way of doing it on Excel: you need to set up the formula yourself. Or use more sophisticated statistical analysis software like IBM SPSS. Why 1.96? There’s a very good reason, but for that you need to look at one of the more in-depth introductions to stats. 

effect sizes

Effect sizes are a really good statistic to know about when you are reading research papers, because they are one of the most commonly reported statistics these days. Also, if you are wanting to compare anything statistically—for instance, whether boys or girls have higher levels of distress when they come into counselling—you’ll want to be giving an effect size.

In fact, there are hundreds and hundreds of different effect size statistics. An effect size is just an indicator of the magnitude of a relationship between two variables. So that might be gender and levels of psychological distress, or it might be the relationship between the number of sessions of art therapy and subsequent ratings of satisfaction. Whatever effect size statistics is used, though, the higher it is the stronger the relationship between two variables.

  • Cohen’s d. The most common form of effect size that you see in the therapy research literature is Cohen’s d, or some variant of it (for instance, ‘Hedges’s g’ or the ‘standardised mean difference’). This is used to indicate the difference between two groups on some variable. For instance, we could use it to indicate the amount of difference in levels of psychological distress for boys and girls coming into counselling, or to indicate how much difference counselling made to young people’s levels of psychological distress as compared with care as usual (which is what we did in our ETHOS study). Cohen’s d is basically the amount of difference between two scores divided by their standard deviations. So, for instance, if boys had a mean level of distress on the YP-CORE of 20, and girls had a mean of 22, and the standard deviation across the two groups was 4.0, then we would have an effect size of 0.5. (This is the difference between 22 and 20 (i.e., 2 points) divided by 4.0). Dividing the raw difference in scores by the standard deviation is important because if, for instance, boys’ and girls’ scores varied very markedly already (i.e., a larger standard deviation), then a difference of 2 points between the two groups would be less meaningful than if the differences in scores were otherwise very small. Typically, when we interpret effect sizes like Cohen’s d:

    • 0.2 = a small effect

    • 0.5 = a medium effect

    • 0.8 = a large effect

    So we could say that there is a medium difference between girls and boys when coming into counselling. In our study of humanistic counselling schools, we found an effect size of 0.25 on YP-CORE scores after 12-weeks between the young people who had counselling and those who didn’t, suggesting that the counselling had a small effect. We can also put a confidence interval around that effect size, for instance ours was 0.03 to 0.47, indicating that we were 95% confident that the true effect of our intervention on young people would lie somewhere between those two figures.

correlational analyses

Correlations are, actually, another form of effect size. But they specifically tell us about the size of relationship between linear variables (i.e., where the scores vary along a numerical scale, like age or YP-CORE scores) rather than between a linear variable and categorical variable (i.e., where there are different types of things, like White vs. BAME, or counselling vs. no counselling).

  • Correlations. These are used to indicate the magnitude of relationship between just two linear variables. It’s a number that ranges from -1 to 0 to +1. A negative correlation indicates that, as one number goes up the other goes down. So, for instance, a correlation of -.8 between age and levels of psychological distress would indicate that, as children get older, their levels of distress go down. A correlation of 0 would indicate that these two variables weren’t related in any way. A positive number would indicate that, as children get older, so they are more distressed. Correlations can be easily calculated on Excel using the function CORREL. Typically, in interpreting correlations

    • 0.1 = a small association

    • 0.3 = a medium association

    • 0.5 = a large association

Tables

If you’ve got lots of different bits of quantitative data (say six or more means/SDs), it’s generally good to present it in a table. Below, for instance, is a table that we used to present data from our ETHOS study about young people who had school-based humanistic counselling plus pastoral care as usual (SBHC plus PCAU group) and those who had pastoral care as usual alone (PCAU group).

In our text, we also gave a narrative account of the main details here (for instance, how many females and males) but the table allowed us to present a lot of detail that we didn’t need to talk the reader through. Generally, tables are a better way of presenting the data than figures, such as graphs, because they can more precisely convey the information to a reader (for instance, a reader won’t know the decimal points from a graph). Just to add, if you are doing a table of participant demographics, the format above is a pretty good way to do it, with different characteristics listed in the left hand column, grouped under subheadings (like ‘Disability’). That works even when there’s just one group, and is generally better than trying to do different characteristics across the top.

Graphs

…But graphs do look prettier, and sometimes they can communicate key relationships between variables that a table or narrative might not. For instance, below is a graph showing our ETHOS results that gives a pretty clear picture of how our two groups changed on our key outcome measure of psychological distress over time. This gives a very immediate representation of what our findings were, and can be particularly useful when conveying results to a lay audience. However, for an academic audience, graphs can be relatively imprecise: if you wanted to know the exact scores, you’d need to get a ruler out! So use graphs sparingly in your own reports and only when they really convey something that can’t be said in a table. And I’d generally say NAAPC (nearly always avoid pie charts): you can get some lovely colours in them, but they take up lots of space and don’t tend to communicate that much information.

Main outcomes from the ETHOS study

Inferential Statistics

Basic principles

So now we come on to the second main type of quantitative analysis: inferential statistics. This is where we use numbers to test hypotheses: that is, we’re not just describing the data here but trying to test particular beliefs and assumptions. Inferential statistics are notoriously difficult to get your head around, so let’s start by taking a step back and thinking about the problem that they’re trying to solve.

Let’s say we find that, after 10 weeks of dramatherapy, older adults have a mean score of 15 on the PHQ-9 measure of depression, while those who didn’t participate in dramatherapy have a mean score of 16. Higher scores on the PHQ-9 mean more depression, but is this difference really meaningful? What, for instance, if those who had dramatherapy had mean scores of 15.9, as opposed to 16.0 for those without—what would we make of that? The problem is, there’s always going to be some random variations between groups—for instance, one might start off with more depressed people—so any small differences between outcomes might be due to that. So how can we say, for instance, whether a difference of 0.1 points between groups is meaningful, or a difference of 1 point, or a difference of 10 points? What we’re asking here, essentially, is whether the differences we have found between our samples are just a result of random variations, or whether they reflect real differences in the population means. That is, in the real world, overall, does dramatherapy actually bring about more reductions in depression for older adults?

So here’s what we can do, and it’s a pretty brilliant—albeit somewhat quirky, on first hearing—solution to this problem. Let’s take our difference of 1 point on the PHQ-9 between our dramatherapy and our no dramatherapy groups. Now, we can never say, for sure, whether this 1 point difference does reflect a real population difference/effect, because there’s always the possibility that our results are due to random variations in sampling. But what we can do is to work out the probability that the difference we have found is simply due to random variations in sampling. The way we do this is by saying, ‘If there were no real differences between the two groups (the null hypothesis), how likely is it that we would have got this result?’ For instance, ‘If dramatherapy was not effective at all, how likely is it that we would have got a 1 point difference between the two groups?’ We can work that out basically by looking at the ratio between how much scores tend to vary anyway across people (i.e., the standard deviation), and then how much they vary between the two specific groups. For instance, if we find lots of differences in how older adults score on the PHQ-9 after therapy, and only very small differences between those who had, and did not have, dramatherapy, the likelihood that the mean differences between the two groups would be due to just random variations would be fairly high. The exact method to calculate this ratio is beyond this blog (and Excel too—you generally need proper statistical software), but the key figure that comes out of it all is a probability value, or p-value. So this is a number, from 1.0 downwards, which tells you how likely it is that your results are just due to chance. So you might get a p-value of .27 (which means that there is a 27% likelihood that your results were due to chance) or .001 (which means that there is a 0.1%, one-in-a-thousand, likelihood that it was due to chance).

So what do you do with that? Well, the standard procedure is to set a cut off point and to say that, if our probability-value is less than that, we’ll say that our difference is significant. That cut-off point is typically .05 (i.e., 1-in-20), and sometimes .01 (i.e., 1-in-100). So, essentially, what we do is to see whether the probability of our results coming about by random is 1-in-20 or less and, if it is, we say that we have a significant result. Why 1-in-20? Well, that’s a bit random in itself, but it’s an established norm, and pretty much any paper you see will use that cut off point to assess whether the likelihood is so low that we’re going to say we’ve found a meaningful difference. Note, if we don’t find a p-value of less than 1-in-20 we can’t say that we’ve shown two things are the same. For instance, if our p-value for dramatherapy against no dramatherapy was 0.27, it doesn’t prove that dramatherapy is no more effective than no dramatherapy. It just means that, at this point, we can’t claim that we have found a significant difference.

Statistical tests

There are a large number of statistical tests that you’ll see in the literature, all based on the principles outlined above. That is, they all ways of looking at different sets of data and asking the question, ‘How likely is it that these results came about by chance?’ If it’s less than 1-in-20, then the null hypothesis that the results are just due to random variations is rejected, and a significant finding is claimed. That’s what researchers are looking for; and it’s a bit weird because, as you can see, what we’re trying to do is to disprove something we never really believed in in the first place! It’s all based, though, around the principle that you can only ever disprove things, not prove things—see Karl Popper’s work on falsifiability here.

Some of the most common families of statistical tests you will come across are:

  • T-tests. These are the most simple tests, and compare the means of two groups. This may be ‘between-participants’ (for instance, PHQ-9 scores for people who have dramatherapy, and those who do not have therapy) or ‘within-participants’ (for instance, PHQ-9 scores for people at the start of dramatherapy, and then at the end).

  • Analysis of variance (ANOVAs). These are a family of tests that compare scores across two or more different groups. For instance, the PHQ-9 scores of participants in dramatherapy, counselling, and acupuncture could be compared against each other. Multiple analyses of variance allows you to compare scores on different dimensions, and then also the interactions between the different dimensions. For instance, an experimental study might look at the outcomes of these three different interventions, and then also compare short term and long term formats. Repeated measures analyses of variance combine within- and between-participant analyses: comparing, for instance, changes on the PHQ-9 from start of therapy to end of therapy for clients in dramatherapy, as compared with one or more other interventions.

  • Correlational tests. Correlations (see above), like differences in means, are very rarely exactly 0, so how do we know if they are meaningful or not? Again, we can use statistical testing to generate a p-value, indicating how likely the association we find is due to random chance.

  • Regression analysis. Regression analysis is an extension of correlational testing. It is a way of looking at the relationship between one linear variable (for instance, psychological distress) and a whole host of other linear variables at the same time (for instance, age, income level, psychological mindedness). Categorical variables, like gender or ethnicity, can also be entered into regression analyses by converting them into linear variables (for instance, White becomes a 1 for ‘yes’, and a 0 for ‘no’). So regression analyses allow you to look at the effects of lots of different factors all at once, and to work out which ones are actually predictive of the outcome and which are not. For instance, correlational tests may show that both age and ethnicity are associated with higher levels of distress, but a regression analysis might indicate that, in fact, age effects are cancelled out once ethnicity is accounted for.

  • Chi-squared tests. As we’ve seen, some data, like gender or diagnoses, is primarily categorical: meaning that it exists in different types/clumps, rather than along continua. So if we’re asking a question like, ‘Are there differences in the extent to which boys and girls are diagnosed with ADHD vs depression?’, we can’t use standard linear-based tests, because there’s no outcome variable. Instead, we use something called a chi-squared test, which is specifically aimed at looking at differences across frequency counts.

… And that’s just the beginning. There’s a mind-boggling number of further tests, like structural equation modelling, cross-panel lag analysis, multilevel modelling, and a whole family of non-parametric tests, but hopefully that gives you a rough idea. There all different procedures, but they’re all based around the same principle: How likely is it, given the results that you got, that there is no difference between the groups? If that likelihood is less than 1-in-20, we’re going to say that something ‘significant’ is going on.

Final thoughts

Whether you like stats or not, they’re there in the research, so if you want to know something of what the research says, you do need to have a basic understanding of them. But we don’t need to get into either/or about it. Stats have their strengths and they have their limitations: from a pluralistic standpoint, they tell one (very helpful at times) story, but it’s not the only story that tells us what’s going on.

Stats, to some extent at least, is also changing. When I trained as a psychology undergraduate in the 1980s, for instance, it was all about significant testing. Today, particularly in psychotherapy research, there’s more emphasis on using stats descriptively, in particular effect sizes and confidence intervals. That’s through a recognition that the kind of yes/no answers you get from inferential tests are too binary and too unrepresentative of the real world.

If you’re staring blankly at this blog and thinking, ‘What the hell was that about?’ do let me know in the comments what wasn’t clear and I’ll try and explain it better. I do, I guess, wish the therapy world would love stats a bit more. I guess that’s partly because it’s so important for understanding what’s getting commissioned and funded and making a difference there; but maybe more because I can see, for myself, so much beauty in it. And that doesn’t in any way take away from the beauty of words or language or art or the many, many other ways of knowing. But numbers can also have a very special place there in helping us to understand people and therapy more; and once you’ve got a basic grasp of what they are trying to do, hopefully they’ll feel more like friend than foe.

Acknowledgements

Photo by Mick Haupt on Unsplash

Disclaimer

 The information, materials, opinions or other content (collectively Content) contained in this blog have been prepared for general information purposes. Whilst I’ve endeavoured to ensure the Content is current and accurate, the Content in this blog is not intended to constitute professional advice and should not be relied on or treated as a substitute for specific advice relevant to particular circumstances. That means that I am not responsible for, nor will be liable for any losses incurred as a result of anyone relying on the Content contained in this blog, on this website, or any external internet sites referenced in or linked in this blog.

The ‘Research Mindset’: Some Pointers

After years of supervising—and teaching—Master’s and doctoral research students in counselling, psychotherapy, and counselling psychology, there’s one thing that, I’ve come to believe, is the key to success. It’s hard to describe, but goes something like this….

When you study or research at undergraduate level, it’s all about showing how much you know. You have to convince your assessors that you are ‘up to it’: that you know enough to meet the learning outcomes for that award.

Students often approach Master’s or doctoral research with the same mindset: they want to show how much they know, that they’re doing it the right way, that they understand the process and the content of the research that they are conducting.

For Master’s and doctoral research that is, indeed, still important; but there is also something much more. When you do research at this level, you are moving from being a student to being a teacher. You, now, are the one who knows. And what the academic community, including your examiners, want from you is not so much to test you or to check your knowledge in a particular field, but to learn from you. We’re looking to you to tell us about what you’re discovering because you know more than us. Yes, that’s right. You do (or, at least, will do); and you need to be able to own that authority.

This can be a hard one: ‘Who am I, I’m just a student, how am I supposed to know anything special?’ But, at doctoral (and to some extent Master’s) level, you are, by definition, being asked to make an original and significant contribution at the leading edge of your field. So, to some extent, this shift needs to happen whether you like it or not. You need to be the big person in the room.

Is this about being arrogant? No, of course not. Is it about pretending you know everything? No, not that either. Is it about patronising your supervisors or your examiners? Definitely not, no. What it is about is being confident and secure in your knowledge and feeling that you have something to educate others about—something to even the most senior figures in the field.

Because the reality is, you do. If you’re researching at Master’s or doctoral level, you should be focusing on a question that no-one else, or very few other people, have ever asked. And that does make you the expert. You know more than us. You know more than your supervisors, you know more than your examiners. You know more than other people in the academic field. And what’s really important to recognise is that we want to learn from you. When someone agrees to examine you for your viva, for instance, or when they come to see you present your research at a conference, they’re not thinking, ‘Mm, I’ve always wondered whether [insert your name here] is good enough for a Master’s/doctoral degree’, or, ‘I’ve always thought [insert your name here] is really just pretending to know things, and I’m now going to find out for sure.’ Nope, that’s probably the last things on their minds. Rather, a large part of the reason they’ve agreed to spend two days reading your thesis and then travelling to your university to examine you, etcetera, is because they’re interested in what you’ve discovered and want to find out more. After all (and apologies to the narcissists here) would anyone really want to spend two or more days of their life just checking up on you? In a world where everyone is so furiously busy what people mostly want is to learn, as effectively and efficiently, what you know so that they can inform and develop their own work and ideas. We want to learn from you.

Doing it despite

Of course that can be scary. When we start off learning in any field, we are inevitably novices; and some of us have ‘imposter syndrome’ throughout our careers. That’s totally understandable. But researching at doctoral and Master’s level means being and doing something despite these fears. It means holding, and owning, our knowledge, skills, and expertise. So if you find it difficult to own that teacher role, this might be something useful to take to therapy: ‘Why is it so difficult for me to see myself as an authority here?’ It gets to the very heart of who we are and how we feel about ourselves.

A key to researching and writing

Although this ‘teacher mindset’ is relatively hard to describe; once you can get into it, it can really unlock the research and writing up process. It means you can write with confidence; and with balance, because you know that what you are saying is important, and that people are wanting a serious, reflective, critical commentary from you. And it means that you are likely to avoid some of the pitfalls stemming from a wholly ‘student mindset’. One problem you sometimes see in students’ theses, for instance, is that their Discussion says next to nothing about their own findings—it focuses solely on the research and theory introduced in their Literature Review. Why does that happen? Probably because, to some extent, the student doesn’t really believe that their own findings have much to say: so they just skip over it and back to the ‘important stuff’. Get into that teacher mindset, however, and you’ll find that you naturally take your own findings much more seriously: they’re not just some throw-away bits of data, they’re carefully curated evidence that have a meaning and significance to the wider field of knowledge.

Narrowing down your focus

One key thing in getting to be—and feeling like—the expert is ensuring that the scope of your research is sufficiently narrow. If you take on a massive area, like ‘the effectiveness of therapy’, you’re never going to feel like (or, indeed, be) the leading authority in that area. There’s people who have spent their lifetimes researching this, carried out hundreds of studies, so, of course, you are going to feel less knowledgeable than them. But if you narrow down your focus—for instance, ‘the effectiveness of compassion-focused therapy (CFT) for health anxiety’—then, immediately, the number of leading authorities in the field dramatically reduces. Sure, people might know about the overall effectiveness of CFT more than you, or the processes by which it supports change; but when it comes to CFT for health anxiety, you’re likely to be in a field of one. And that’s when everyone starts to turn to you to discover what you’ve found, because you’re then genuinely contributing to the knowledge-base. So if you’re feeling like you could never ‘hold’ that expert position in your field, it may be worth looking at how broad your field is. You can, I promise, get to that expertise level, but it is very dependent on the breadth of the question you are asking.

Against authority

But is it OK to be an ‘authority’? Perhaps another block to that teacher mindset, for those of us from more humanistic and person-centred orientations, is that we’re wary of taking on too dominant a role: we don’t like to position ourselves as ‘better’ than others. Here, equality, respect, treating the other as like ourselves are all the principal values. Yes, absolutely; but recognising that we know more than others in one particular field isn’t saying we’re better or smarter than others. We can know lots and others can know lots as well; and if we all share our specialist knowledges—and dialogue between them—then we can all make contributions to a better world for all. Equality doesn’t have to mean sameness. Indeed, recognising our own special knowledges—and giving that away to others—can be part of a world that celebrates difference, diversity, and uniqueness for all.

facing the unknowable

To adopt that teacher mindset, you also have to be willing to face the unknowability of a lot of the questions you are asking. At school and at undergraduate level, the questions you were asked had ‘right’ answers—or, at least, your teachers and lecturers told you they did. Multiple choice questions make it clear that there are ‘rights’ and ‘wrongs’. But when you’re leading the field, when you’re at the cutting edge of developments, there’s often not one right way of going forward. You’re ahead now, and you have to decide which path to cut. Should you use IPA or grounded theory? Two or three levels in your multilevel analysis? Well, sorry, but as your supervisors, examiners, and readers it’s very likely that we don’t actually know. We’ve got our own ideas, but what we’re hoping for is that you’ll be able to face those really difficult questions and, in the absent of any certainties, work if out for yourself (in a sensible, informed, and transparent way). And that’s not because we want to provide a non-directive environment to teach you to work these things out for yourself: it’s because we genuinely, really, don’t know.

That’s what doctoral level competences are about: being able to move forward in the face of incomplete knowledge. If you don’t know, it’s almost certainly not because you are incapable or dumb, but because the reality is that no one else knows either: no one has managed to work it out yet. And what we’re hoping for is that you’ll do the work of working it out. There’s so many questions, uncertainties, and unknowns out there; and if you can take one small chunk of this and do some thinking that can contribute to the wider field, you’ll be doing a massive benefit to all of us.

Conclusion

Be serious, then, about your research. You do nothing for yourself, or for the field, if you treat your research as simply an academic exercise that you have to get through—that isn’t ever going to teach anyone about anything. Sorry if that sounds harsh; but be serious about your research in the same way that you would be serious about your work as a therapist. That doesn’t mean not being able to laugh, or joke, or enjoy it along the way; but it does mean having the confidence to believe that you can give something meaningful to the wider world. And if you don’t feel that, take some time to work on it, in the same way that you would work on your insecurities as a therapist (in research supervisor, for instance, or in therapy, or on your course). Get to a position where, in transactional analysis terms, you’re an adult: where you’re able to own your strengths and your abilities to contribute, as well as your limitations. You have so much to offer.


Acknowledgements

Photo by Ben White on Unsplash

Disclaimer

 The information, materials, opinions or other content (collectively Content) contained in this blog have been prepared for general information purposes. Whilst I’ve endeavoured to ensure the Content is current and accurate, the Content in this blog is not intended to constitute professional advice and should not be relied on or treated as a substitute for specific advice relevant to particular circumstances. That means that I am not responsible for, nor will be liable for any losses incurred as a result of anyone relying on the Content contained in this blog, on this website, or any external internet sites referenced in or linked in this blog.

Overview of the Thesis: Some Pointers

OVERVIEW OF THE THESIS: SOME POINTERS

 

A good thesis is like a journey of discovery: think Odyssey, Lancelot, Charlie and the Chocolate Factory. You’ve set out to find an answer to a question (or some answers to some questions), and each section of your thesis is a stage on that journey:

  • Introduction: why this question is of value

  • Literature Review: how other people have answered it

  • Method: how you will try and answer it

  • Results: what you have found out

  • Discussion: what your findings mean (particularly in relation to previous findings).

To that you can add:

  • Title: Concise statement of your research question/enquiry

  • Abstract: Summary of all the sections in your thesis

  • Conclusion (after Discussion): A summary of what you have found and any outstanding issues.

Each of these sections should be logically linked, so that, if all is as it should be, you should be able to reduce your dissertation down to a single, coherent narrative of not more than a paragraph or so (your abstract, effectively). Below is an example:

The Benefits and Limitations of Using the Two-Chair Technique in Person-Centred Therapy: An Interpretative Phenomenological Analysis

Understanding the benefits and limitations of the two-chair technique in person-centred therapy is important because a number of person-centred approaches, such as person-centred experiential therapy for depression, are moving towards their use [Introduction]. Greenberg has shown that clients can experience the two-chair technique as beneficial, but these findings are primarily quantitative and there is little data on why clients might experience this technique as helpful—or unhelpful—per se [Literature Review]. For this reason, I carried out a series of semi-structured interviews with ten clients in person-centred therapy who engaged with the two-chair technique to find out their views on it. I recruited these clients through social media. Their interview data was analysed through interpretative phenomenological analysis [Method]. In terms of benefits, clients said that the two-chair technique had helped them express feelings that they found difficult to express otherwise. It also helped them identify different aspects of themselves, and helped them feel closer to their therapist. On the other hand, for some clients the two-chair technique had interfered with the therapeutic alliance. The main reason for this was that it made them feel embarrassed, and in one case the client actually left counselling as a result [Results]. It would seem, then, that Greenberg and others are right that the two-chair technique can be very useful as an adjunct to person-centred therapy, however there may be some contraindications of this technique [Discussion].

Have a go at summarising your research in this way, whatever stage you are up to in your write-up (even if you haven’t started). And if you don’t have results yet, have a go at just imagining what results to your question might be: this isn’t about prefiguring your Results, but about getting some sense of what might be meaningful answers to your questions.

A key issue to focus on here is whether all the different sections of your thesis are aligned. That is, are they all oriented around the same question(s) or, rather, are they actually asking and answering different questions? You may find, for instance, that the Literature Review you’re planning doesn’t really answer the questions you’re posing in your title; or that your Results are answering a different set of questions to the ones you reviewed the literature on. These differences may be subtle (for instance, your Literature Review might focus on what clients find helpful in therapy whereas your Results reveal what they experience in therapy), but any minor differences can become magnified as your research and write-up progresses. And such differences can really ‘do your head in’, because you can then get into a terrible muddle about what it is you are asking and answering. So make sure you have a set of really clearly defined questions (see blog here), and then ensure all the sections of your thesis revolve around that.

One way of developing that alignment is by taking your Title or Literature Review and asking yourself, ‘What might be meaningful answers to the questions that I am posing and addressing here?’ Then you can look at your Results and see if they do, indeed, match the questions being asked. Or you can do the same process backwards: Take your Results and ask yourself, ‘What kind of questions are these answering?’ If your Results are providing answers to questions that, in fact, you never asked in the first place, you know you need to do some work on re-aligning your thesis.

The ‘Cafe Test’

Here’s another way of checking the coherence of your research study. Imagine that you are sitting in a cafe chatting to a friend. They are really interested in your work, and they ask you what you have discovered in relation to your research question. What would you say to them? How would you answer that question in a simple, non-jargon way—that gave them a genuine, meaningful, interesting answers to their question. You might want to actually try that with a friend and record your answer – because often that is the most succinct summary of your research and your findings that you’ll give. And if you don’t have results yet, again, try it anyway, and just imagine what kind of results you might have. Note if you are finding it difficult to respond; and also note if the response you are giving in the Café Test is very different from what you have written—or are planning to write—in your thesis itself. Is it, for instance, that you have written a lot in your thesis that isn’t actually that related to your research question? Or perhaps emphasised answers that, in more everyday conversation, aren’t actually that interesting. Try, wherever possible, to align your Results and Discussions to what you would, genuinely, describe as meaningful in an everyday conversation—although, of course, the language and structure needs to be more formal in your write-up. That’s the most exciting bit of what you’ve found, so ‘big it up’: make sure it’s at the heart of what you are communicating to your academic audience.

Acknowledgements

Photo by Handy Wicaksono on Unsplash

Disclaimer

 The information, materials, opinions or other content (collectively Content) contained in this blog have been prepared for general information purposes. Whilst I’ve endeavoured to ensure the Content is current and accurate, the Content in this blog is not intended to constitute professional advice and should not be relied on or treated as a substitute for specific advice relevant to particular circumstances. That means that I am not responsible for, nor will be liable for any losses incurred as a result of anyone relying on the Content contained in this blog, on this website, or any external internet sites referenced in or linked in this blog.

Research Aims and Questions: Some Pointers

Your aims are the beating heart of your research project, and your write-up. Whether you are conducting an exploratory study or a hypothesis testing one, whether qualitative of quantitative, you are trying to do something in your research, and specifying what that doing is is the key that holds your project together.

Wherever you are in a research project, try specifying exactly what your aims for it are, for instance:

In this project… I am trying to discover how clients’ experience preference work

In this project… I am trying to find out if school counselling is effective

In this project… I am trying to assess the psychometric properties of the Goals Form

In research, the aim is to always find something out, so it’s always possible to also reframe your aims as a question:

How do clients’ experience preference work?

Is school counselling is effective?

What are the psychometric properties of the Goals Form?

Framing it either way is fine. But it’s essential that your aims and your questions match, and it’s generally helpful to be aware of both forms as you progress through your research.

If you’re struggling to articulate the aims of your research, ask a friend or peer to ‘interview’ you about it. They can ask you questions like:

  • ‘What are you trying to find out?’

  • ‘What’s the question that you are asking?’

  • ‘What do you want to know that isn’t known up to this point?’

  • ‘What kind of outcome to this project would tell you it’s been a successful one?’

Trying to articulate your research aims/questions isn’t always easy, and it’s generally an iterative process: one that develops as your research progresses. Sometimes, it’s a bit like an ‘unclear felt sense’ (from the world of focusing): you kind of ‘know’ what the aim is, but can’t quite put it into words. It’s on the tip of your tongue. That’s why it can be helpful to have a colleague interview you about it so you can try and get it more clearly stated.

Another way into this would be to ask yourself (or discuss with peers):

  • ‘What might be meaningful findings from my project?’

For instance, with the research questions above, meaningful findings might be that ‘clients find it irritating to be asked about their preferences’, or that ‘the Goals Form has good reliability but poor validity?’ Of course, you don’t want to pre-empt your answers, but just seeing if there are potential meaningful answers is a good way of checking whether your question makes sense and is worth asking. If you find, for instance, that you just can’t envisage a meaningful answer, or that the only meaningful answers are ones that you already know about, it may mean that you need to rethink your research question(s). There needs to be, at least potentially, the possibility of something interesting coming out of your study.

You may have just one aim, you may have more than one aims. A few aims is fine, but make sure there aren’t too many, and make sure you’re clear about what they are and how they differ. Disentangling your aims/research questions can be complex, but it’s essential in a research project to be able to do that: so that you and whoever reads your research knows what it’s all about, and what your contribution to knowledge might be.

If you find it difficult to articulate your aim(s), it may be that, at the end of the day, you’re not really sure what your research is about. That’s fine: it’s a place that many of us get to, particularly if our research has gone through various twists and turns. So it’s not something to beat yourself up about, but it is something to reflect on and see if you can re-specify what it is, now, that you’re trying to do and ask, so that you can be clear. This may mean turning away from some of the things you’ve been interested in, or some of the questions that you were originally asking. It can be sad to let go of aims and questions; but it’s generally essential in ensuring you’ve got a nice, clear, focused project—not one where you’re going to be lost in a forest of questions and confusion.

If you specify your aims but can’t rephrase them as questions that’s also worth noting. That may be an indicator that really what you are trying to do is to prove something, rather than conducting a genuine inquiry. For instance, you may find that your aim is, ‘to show that people living in poverty cannot access counselling?’ or ‘to establish that female clients prefer self-disclosure to male clients’. If that’s the case, try and find a way of re-framing your research in terms of an open question(s): one(s) that you genuinely don’t know the answer to. It’s so much more powerful, interesting, and meaningful to conduct research that way. Indeed, if you’re struggling to articulate your research question, one really valuable question to ask yourself is:

  • ‘What is the question that I genuinely don’t know the answer to?’

And ‘genuinely’ here does mean genuinely. If you’re pretending to yourself that you don’t know something so that you can show it anyway, then that’s likely to become evident when you write up your research. So really see if you can find a question that you genuinely, really genuinely, can’t answer at this point—but one that you would really love to be able to. That’s a fantastic place to start research from.

Once you’ve got your beating heart, write it up on a stick it note and put it on your wall somewhere or put it on your screensaver. Keep it in mind all the time: the aims of your research and the questions you’re asking. When you’re interviewing your participants, when you’re doing your analysis… keep coming back to it again and again. It’ll keep you focused, it’ll mean that you keep on track, and it’ll keep you with a clear sense of where it is you want to go and what you are trying to do.

If you deviate, that’s fine, we all do that. Just like in meditation, notice you’re moving on, then try and bring yourself back. Or, if you really can’t bring yourself back to your aims/questions, then it may be that they need to change. That’s fine in a research project and it does happen but, again, be clear and specific about what the aims and questions are changing to, and make sure that the rest of your project is then aligned with those new directions. What you don’t want, for instance, is a Literature Review asking one set of questions, and then a Results section that answers an entirely (or even slightly) different set of aims.

And when you write up your thesis or research paper, start with your aim(s)/question(s). Often people put them towards the end of the Literature Review (i.e., just before the Methods section), but you can also put them earlier on in your Introduction. Write them down just as they have been formulated as you’ve progressed: clear, succinct, a line or two for each. If there’s more than one, write them down clearly as separate aims/questions. You probably don’t need to give them in both formats and you could use different formats in different places: for instance, they could be stated as aims in your Abstract and Introduction, then as questions just before your Methods section.

Once you’ve got those aims/questions stated, you can build all the other parts of the research and write-up around it. For instance:

  • Literature Review section: You can structure this by the questions you’re asking, with different sections looking at what we know, so far, in relation to each question.

  • Interview questions: In most instances, the questions you ask your participants should match, pretty much exactly, your overarching research questions. So if you are interested in how clients experience preference work… ask them. No need to faff about with indirect or tangential interview questions: just go into the heart of what you really want to know, and have a rich, complex, multifaceted dialogue about it.

  • Results section: Whether qualitative or quantitative, you can present your findings by research question: So what did you find in relation to question a, then in relation to question b, etc.

  • Discussion section: This, too, can be structured by research question—though I would tend to do this in the Discussion or in the Results (not both), so that the sections don’t overlap too much with each other.

  • Limitations: Don’t just say what’s good or bad about your research: say how the answer you got to your questions might have been biased by particular factors, and what that might mean.

  • Abstract: When you come on to write this, make sure your aims/questions are clearly stated, and then clear answers to each question are given.

Being clear about your research aims and questions, and focusing your research around them, may seem obvious. It may also seem pedantic or overly-explicit. But it’s key to creating a coherent, focused research project that—as required at master’s or doctoral level—makes a contribution to knowledge. It can be hard to do; but working out, for yourself, what you are trying to do and ask is a key element of the research process. Research isn’t just a question of mucking in, generating data, and leaving it to your reader (or your assessor) to work out what it all means. You need to do that: to guide the reader from question(s) to answer(s), and to help them see how the world is a better-understood place (even if it’s just a little better understood) for what you have done.

Acknowledgements

Photo by Bart LaRue on Unsplash

DISCLAIMER

The information, materials, opinions or other content (collectively Content) contained in this blog have been prepared for general information purposes. Whilst I’ve endeavoured to ensure the Content is current and accurate, the Content in this blog is not intended to constitute professional advice and should not be relied on or treated as a substitute for specific advice relevant to particular circumstances. That means that I am not responsible for, nor will be liable for any losses incurred as a result of anyone relying on the Content contained in this blog, on this website, or any external internet sites referenced in or linked in this blog.

How to (almost) Fail a PhD: A Personal Account

The year, 1996, didn’t start well. My then-partner and I went to Spain, with three friends, for a Christmas break. For some reason we thought it would be shining hot. As it turned out, we spent a week in a wet, damp bungalow in the middle of nowhere. The main thing I remember were the Spanish tortillas on the few days we got out—wet and damp as well, with burnt soggy potatoes at the bottom.

My PhD viva was on Friday the 6th Jan—25 years from the publication of this post (more details on what a PhD viva is are available here). I’d read through my thesis a few times and felt fairly well-prepared. It was a somewhat unusual topic, Facilitating the expression of subpersonalities through the use of masks: An exploratory study. Basically, during my undergraduate studies I’d gone to a mask workshop run by a friend of mine at Oxford University and been amazed at the power of masks to bring out different ‘sides’ of my self (or ‘subpersonalities’). I researched it further for an undergraduate paper and then, in the early 1990s, applied to Sussex University to do a PhD on the topic. Basically, I wasn’t sure what I wanted to do as a career—either media (TV, journalism) or academia—and, as I couldn’t find a way in to media work, I thought I’d do the latter, particularly when I was awarded a grant from Sussex University to support me. That’s when I also started counselling training: I thought I better to do something practical alongside the PhD.

The internal examiner for the viva was a tutor of mine from my undergraduate days and someone who I knew fairly well. The external examiner was an academic in humanistic psychology I didn’t know much about, but had read a couple of her books and they seemed interesting. The three of us sat that Friday in the internal examiner’s office: dark and small, with his bike leaning against the bookshelves.

I remember more about after the viva than the viva itself. But the questions came quickly and they felt pretty intense from the start. ‘Why was I writing about subpersonalities?’ ‘What evidence was there for them?’ ‘What made me think they were a legitimate basis for a PhD?’ ‘Why was I so dependent on the work of John Rowan, what about my own thoughts?’ I answered the questions as best I could, wondering if that was how vivas were supposed to be—anxious that, perhaps, this was more critical than normal. After about 90 minutes I was asked to leave and sat in the Department common room—somewhere I’d spent many hours as an undergraduate socialising and relaxing in. I felt a rising anxiety from the pit of my stomach. I’d done my best, but something felt wrong. One of my other undergraduate tutors passed by and asked me how things had gone. He said he was sure it would all be fine: no one got failed for their viva. I wasn’t so sure.

Called back in the darkened room, like a death sentence. They had, indeed, decided to fail the thesis. Well, not quite fail it, but they were proposing that I resubmit for an MPhil: the next to lowest outcome. The main thing I remember was crying. I think it was an armchair I was sitting in, in a corner of the room. Sobbing away. Couldn’t believe it, even though I’d felt it coming. I went to see my supervisor and told him the news. Then I walked and walked and walked to a nearby village. Bought some cigarettes for the first time in years, rang my closest friend from a red call box and just smoked and smoked. There was nothing else I could do.

I came back to campus and went to see my supervisor again. He said that the examiners had decided that, in fact, I could have another chance to resubmit for a PhD: one outcome higher. But it would require a complete rewrite—four years’ work down the drain!

I met my partner at our house near Brighton station. Then we went to the pub. A few pints and I felt better, but I knew it was just temporary. Back home, as the alcohol wore off, the reality of the situation smashed back in my face. And so many questions: ‘Why had I failed?’ ‘Why had my supervisor said to me, just the day before the viva, that the work was “excellent”?’ ‘Were they ever likely to pass it even if I did spend the next three years rewriting?’ More than anything, I just didn’t understand what was wrong with the work, why they had failed it. The examiners obviously, clearly, really didn’t like it. But why?

That weekend was probably the worst of my life. I hardly slept the Friday night, just terrible feelings of anxiety and worry. Thinking over and over again what had gone wrong. A few hours sleep, then pub the next day and again some temporary relief. Then walking, walking, walking with my partner—along Brighton seafront—trying to make sense of things and work out ways forward. A game of pool in a pub in Hove. Slow walk back along the Western Road. I bought some aftershave at a chemist in Seven Dials that was my favourite for many years. Back home in the silence and the pain of it all. Moments alone were the worst, when my partner went to sleep. Several serious suicide attempts over the next few days. I won’t go into details, but suffice to say that it was just the terror of the pain, and the thought of having—and meeting my—children in the future, that held me back.

It wasn’t just failing my thesis. It was where I was in life. Basically, I was 30, had been struggling for years to work out what I wanted to do. Had been watching so many family members and friends succeed in their careers. I felt like I was going nowhere. The one thing I had was this PhD and the possibility of being an academic, and now even that was in tatters. It was the last closed door in a series of closed doors. The last possibilities I’d been hanging out for.

One of the worst things was that I had to run seminars for the psychology undergraduate students the next week. I felt so totally and utterly ashamed: surely everyone would know about my failure, and then how could they possibly take anything I said seriously? I drove in that Wednesday, facilitated the class as best I could. It didn’t help that the internal examiner was the module coordinator. I spoke to him as well on the phone on the Monday. He was sorry to hear I was feeling so awful. He tried to explain what had happened but it just didn’t make any sense. More questions, not less.

I was teaching psychology at Brighton University as well at that time, and was so grateful that the programme coordinator there didn’t seem to flinch when she heard the news. She still trusted me, let me continue my teaching. In fact, that summer, when she moved on, I was offered her job, and started in a more permanent position at Brighton University.

Something had already seemed to turn, though, before that time. I felt a bit better by April. I had a new supervisor now (one of the conditions for me being allowed to resubmit): a professor from my undergraduate years that I really trusted. He was down-to-earth, grounded, gave me hope. But it was a whole new thesis, and three more years before I finally completed.

What Went Wrong?

So why had things gone so badly wrong? Had my supervisor let me down, was it that the examiners had been unfair, or had I just done a really poor piece of work? It took me months, maybe even years, to work out. But now I’d understand it something like this: When I started the work, I was doing it in the field of cultural studies. It was about masks, and with a fairly relaxed design: I was drawing on literature, ethnography, drama therapy. There was no stringent method, but that seemed fine for that field of study and others who wrote a thesis in a similar way had done fine. But then, about halfway through my programme, we’d shifted my registration to Psychology. My supervisor, I think rightly, wanted me to come out with a doctorate in psychology so that I could use that if I wanted to go into psychology as a profession—for teaching or clinically. But the problem was, the focus or content of my thesis hadn’t really changed. So my examiners, who were fairly classical psychologists, thought the whole thing was just off the wall. Far too a-methodological, no real use of systematic methods or analysis. As a psychology thesis yes, they were right, it didn’t meet expected standards. But I had no idea what those standards were. And somehow my supervisor had never seen that coming. And I guess I hadn’t too. There were warning signs. For instance, I presented at my psychology department’s seminar series and I could see that they weren’t too taken by being asked to wear masks and to move around in them, but I hadn’t wanted to see the problems. And I should have pushed harder for a second supervisor. I did ask, and it was discussed, but I let it go and thought it would all be OK.

What’s the Learning?

I guess, as with all awful things, there was a lot of learning. That experience has stayed with me throughout my life. I still go back to that pub by Brighton station every so often to sit and reflect and thank something or someone for, in the end, making things OK. And I’d do that again tonight if it wasn’t for COVID. Somehow, amazingly, within ten years of that viva I was a professor of counselling at a prestigious university in Scotland: something, sitting back there in 1996, I could never have even hoped for. When I go back to the pub, I kind of ‘talk’ to my 1996 self and tell him that things are going to be OK in the end, and to hang on in there. And it’s nice, in some ways, to have that chat with him and reflect on where things ended up. He’d have been so happy and relieved.

As a Student

One thing that I really did that was wrong was to isolate myself away from any academic community while I was working on my PhD. I never went to conference, or engaged with departmental seminars, or submitted to journals. And just the one time I did present, as above, I didn’t stay open to how people were responding. I was in my own little bubble, and that wasn’t shattered until my actual viva. I think I did that because I was scared: worried that others wouldn’t be that interested in my work or feel it was good enough. But I made the classic mistake of avoiding, rather than facing up to, the thing I was afraid of.

As a Supervisor

I really try and be straight with my students if I think there’s problems. If I don’t think the work is at the right level, I’ll do my best to say it. Much better they hear it from me than from their examiners.

And when it comes to choosing an examiner for a student, I do think about the importance of ‘alignment’. This is not about finding someone who will simply waive the thesis through; but finding someone who has some of the same basic assumptions and expectations as the student and the supervision team. Most psychologists would probably fail a cultural studies PhD if it was submitted as psychology. And, similarly, I imagine that many cultural studies academics would fail a psychology PhD for reasons—like lack of epistemological, cultural, and personal reflexivity—that traditional psychologists might never consider. So there’s a reality that, in the academic world, there’s lots of different sets of expectations and assumptions; and it seems essential to me that students are assessed in terms of what they are trying—and supported—to do.

These days, most universities (certainly Roehampton) have a minimum of two supervisors for doctoral work, and that’s absolutely key to ensuring that it’s not dependent on just one academic’s views. We do our best, but our blind spots are, by definition, blind spots. Really getting an honest second opinion on student’s work—triangulation—makes it much less likely that things will go off track.

As an Examiner

I’m still angry at my examiners. Fair enough, they didn’t like the work and didn’t think it was at doctoral standards. But, they were so critical, so personal about the problems in the thesis. The external examiner, in particular, felt just ‘mean’ at times. When my new supervisor and I wrote to her, while I was revising, just to check I was along the right lines, she wrote a response that felt so demotivating and unclear. It just wasn’t needed. So when I’m a doctoral examiner now, even if I feel more work needs to be done, I try and do it supportively and warmly—with kindness, sensitivity, and empathy.

There’s also something about acknowledging the multiplicity of perspectives on things. As an examiner, I have to give my perspective on what I think is doctoral standard, I can’t ever be entirely objective; but I can acknowledge it as my perspective. You can criticise something without criticising the person behind it.

As a Person

I guess one of the best things that came out of this whole period of my life is that I’ve never taken my job for granted. I feel incredibly privileged to have had a chance to work and teach: just seeing students, writing emails—it’s amazing to have this role and this opportunity with others. I still, deep down, don’t believe that I would/will ever have it.

I guess the downside of this, which has not been so great for relationships and, perhaps, as a father, is that I’m still so focused on work. If I don’t do a set number of hours each day, I start to feel almost shaky and that I’m letting work down. I’ve worked, maybe, 55 hour per weeks for the last twenty or so years. Rarely taken my full annual leave. And that’s, in part I’m sure, because I’m still haunted by the ghost of that experience. My 1996 self still regularly tells me ‘You’ll never have a job’, ‘You’ll never be part of a work team,’ ‘You’ll always be a failure and outside of things.’

Something else at the edges of my awareness: when I look back, I realise how much I had to contribute at that time. So much passion, energy, commitment. I really wanted to make a difference. And it was so, so hard to—not just with the PhD but as a young person struggling through their 20s who didn’t quite fit into the social structures. And it makes me think about how much of that energy gets wasted in our young people: so much passion, drive, and creativity that is blocked, that doesn’t have an outlet. It’s such a burning frustration for those young people, and such a waste for our society as a whole.

Concluding Thoughts

I still feel shaky, and then some relief, reflecting on this time. I’ve never written about it before and perhaps there’s still more to process in therapy. Just that sheer, pounding, devastating sense of failure and shame. But there’s also something profoundly uplifting about it. How you can be right at the very bottom of things, utterly hopeless, but if you stick with things and keep going despite then it can get better and amazing things can happen. I’d love to say ‘trust the process’ or that, in some way, that failure led to subsequent successes; but in many ways I think I was just incredibly, incredibly lucky that things worked out ok. Part of me, maybe that 1996 part, believes (or, perhaps, knows) I could still be struggling away. And I do feel like I’ve been amazingly lucky and blessed in my career and in my life: more than anything, four beautiful, gorgeous children.

Out of the storm, chaos, and anguish of life, there’s still the possibility of some incredible things emerging. Things can change. Even when we’ve totally given up on hope, hope and possibility may still hold out for us.


Acknowledgements

I am deeply indebted to Helen Cruthers, James Sanderson, and the friends and colleagues who helped me through that time in my life.

Very special thanks to Christine Aubrey—I will always be so grateful.

Thanks also to Yannis Fronimos for feedback and encouragement on this article.

A condensed version of this article was published in the BACP publication Therapy Today and can be downloaded here. Thanks to Sally Brown for her superb editing and condensing of the post.

Recruiting Research Participants: Some Pointers

Participant recruitment… it’s the make-or-break of many a research project, so it’s surprising that it’s not addressed more in the literature. It’s as if, once you’ve chosen your research questions, decided on your methodology, and obtained ethical approval, you just close your eyes and, as if with a sprinkling of fairy dust, your data appears….

If only! Truth is, finding people to take part in your study is often the hardest, and most gruelling—emotionally as well as physically—part of your research. And difficulties over recruitment are one of the main reasons why people have to extend their research projects—sometimes by years. So if you want to make sure your research project is a successful one, planning for recruitment is something you need to take seriously, right from the very start.

Who’s busy too?

Why is it so hard to recruit for your study? Well, first, there’s a good chance that most people aren’t really going to want to do it. Sorry. That’s not to say that they’ll be critical of your research or think it’s pointless. It’s just that so many of us are so incredibly busy these days. Think of how you feel when you see an email or a Facebook notice inviting you to take part in some research. ‘Mm, looks kind of interesting,’ but with a hundred plates already spinning in your life, do you really want to take on one thing more? Have you got a spare hour and time to read information sheets and fill in consent forms. With your kids screaming in the next room or your partner who’s just put dinner on the table! So however fascinated you might be with your own research topic; remember that other people are ‘outside’ of your head rather than ‘inside’ of it: caught up in their own world of worries, tasks, and goals.

Touching on sensitivities

Prospective participants may also be reluctant to take part because what we’re inviting them to do is hard, emotionally as well as cognitively. Counselling psychologist Jasmine Childs-Fegredo says:  

In a qualitative study for example, you might be asking people quite personal, and possibly even potentially distressing questions, about their past experiences. For a quantitative study, you might be asking participants to undertake a task or complete a survey which they might be worried about finding difficult in some way. 

Jasmine goes on to add:

So you need to approach your recruitment strategy really sensitively: from the wording in a recruitment poster to the emails you might send out to participants. Use warmth, empathy, and be professional; enabling people to feel safe and thereby prepared to be in a relationship with you as their researcher.

Rosie Rizq, a former Professor at the University of Roehampton, makes a similar point when she emphasises the importance of developing a collaborative relationship with (prospective) participants right from the very start:

In my experience, many trainees tend to revert to a 'helicopter in/helicopter out' approach to their participants, rather than approaching potential participants with a collaborative mindset that may require particular sensitivity and thoughtfulness. I guess I'm seeing participant recruitment as part and parcel of a wider mindset and epistemology required for projects that might involve highly personal, painful, or sensitive material. Some participants require very careful handling indeed prior to any agreement being signed off; they need to feel a sense of confidence in the researcher and that their material will be treated carefully and respectfully before, during and, most importantly, after interviews.

So recruitment is not something we do to prospective participants; it’s a way, if you like, of initiating a relationship. And as with the start of all relationships it needs to be done with care, sensitivity, and attention.

Awkward!

But there’s another, third, reason why recruitment can be so difficult: because, for us as researchers too, it can just feel so incredibly awkward. Maybe this is just for the introverts amongst us, but I remember, when running psychology experiments for my PhD, just how excruciatingly embarrassed I felt asking people to take part—I wanted to die inside. Perhaps it’s a fear of rejection; perhaps an anxiety about receiving without giving—kind of like ‘pleading’ people to do something. I remember, as a kid, having the same feeling when I’d spent all my money in the arcades and had to beg random strangers for the tube fare home (in fact, I found it so torturous that my best friend, James, always ended up doing the begging for us!). Seriously, though, that awkwardness can create real obstacles towards successfully recruiting for—and completing—a study. You can’t do recruitment if you’re metaphorically hiding behind a wall somewhere, secretly hoping that no-one will notice you.

Be Proactive

So, as a general principal for successful recruitment, a key thing is to be proactive. This doesn’t mean being pushy, demanding, or nagging people when they’ve clearly had enough of you. But it does mean taking active steps to make recruitment happen, being on top of it, and pushing through—where appropriate—your own embarrassment or awkwardness barrier. Remember… there’s no fairy dust. Recruitment will not happen to you. Over the years, I’ve just seen too many research projects fail, or severely stall, because researchers have sat back and waited for participants to arrive rather than actively seeking them out.

Planning recruitment from the start

All this means that a recruitment strategy should be built into your research project from the very start, not an add on once you’re through ethical approval. And if you can’t conceive of viable ways to recruit people to a particular study, it may well be that you need to do something else—there’s no point having a brilliantly designed study if no one is actually able or willing to do it. Those strategies need to be concrete, realistic, and well thought out; and ideally tried-and-tested. Has an organisation, for instance, given initial indications that they would be willing to support recruitment, or have other projects used similar strategies to good effect? Remember, too, that of the many people who are potentially available to take part, most won’t. So if you’re planning, for instance, to interview clients at a service that sees 100 people per year, you might have to assume that at least 80% or so of them won’t be interested—and then of those 20 remaining some will drop out before, or at, the interview. So is that going to leave you with enough participants? However many people you think are going to want to take part in your research, chances are, the final numbers will be even less!

Challenging groups

There’s no doubt that some groups of participants are more difficult to recruit than others. Practitioners are often the easiest to recruit, clients more challenging, and then some groups of clients (for instance, those in prison) next-to-impossible unless you have some specific ‘ins’. The challenges of recruitment with a particular group, however, do need to be weighed up against the value of what research with them will accomplish. So, for instance, although clients may be more difficult to recruit than therapists, they can give much more valuable answers to particular research questions (for instance, ‘How do clients experience therapists’ self-disclosures?’).

If you are planning to conduct research in England which potentially involves ‘research participants identified in the context of, or in connection with, their past or present use of NHS services’ (either as practitioners or clients), you are likely to need NHS REC approval. This can be a time consuming process (3-6 months or more) and one that should be built into any research timeline. Again, though, the value of conducting research with such clients may outweigh the additional demands.

Have a written recruitment plan

A really systematic way to think recruitment through is with a written recruitment plan. This can be done on software like Word or Excel and, in most cases, is something you should be detailing in your ethics submission. List each of the different strategies/channels you’re going to use for recruitment (for instance, Facebook posts, Twitter posts, emails, approaching voluntary organisations), what you’re going to say, when you’re going to do it, and any other relevant details. You can then use that to track recruitment once you’ve started. Are you hitting the timelines you set for your different strategies, and what kinds of responses are you getting? If strategies don’t seem to be successful, strike them out and, where relevant, develop others (but don’t forget that those new strategies might need ethical approval). And, at the end of it all, you can present that plan in the appendix of your thesis: an ‘audit trail’ evidencing how thorough and committed you were in the recruitment process.

Where to recruit from

There’s many different strategies through which you can try and recruit participants, and generally I’d say ‘over do it’ rather than ‘under do it’. That is, given the challenges of finding participants, explore and identify a wider range of strategies than you may actually need, rather than cautiously and conservatively just choosing one or two.

A good starting point is to think where the participants you are looking for may be most likely to ‘congregate’. For instance, do they tend to frequent particular locations (for instance, hospital waiting rooms), or particular online sites (for instance, Reddit ‘subreddits’).

There are numerous places through which you can recruit participants. Some of the most commonly used are:

  • Social media:

    • Facebook: personal pages, or on the many counselling/psychotherapy groups

    • Twitter: use @usernames to add it to the Twitter feeds of organisations like @BACP and @BPSOfficial, or #hashtags (for instance, #counselling)

    • LinkedIn (widely used by professionals)

    • WhatsApp groups

    • Reddit: again, think about groups (‘subreddits’) that may be specific to the topic you’re focusing on.

  • Professional counselling associations (e.g., BACP, UKCP, BPS, BPS divisions): website notice boards, magazines, or research networks

  • Service user groups, networks, and charities—both national and local—like MIND or Triumph Over Phobia (see listing here)

  • Email contacts

  • Websites: personal, university

  • Blogs: for instance, write a blog about your planned research, or the background literature to it, for a relevant site; with a link if people are interested in following up

  • Students: there may be a system, for instance, for psychology undergraduates to participate in research

  • Conference attendees

  • Physical notice boards (for instance, at universities or GP surgeries)—although, in my experience, they are not a particularly fruitful method of recruitment

  • ‘Snowball sampling’: asking your participants to recommend further participants

  • Online recruitment sites (‘mechanical Turks’), like Prolific, where you pay people to complete your survey (this is mainly for large scale, quantitative (and funded!) studies).

Don’t forget that, if your research is conducted digitally (e.g., video conferencing interviews or a web-based survey), you might want to consider recruiting internationally. For instance, if you were looking at the experiences of clients with a particular condition (say counselling for sight loss), you might want to approach service user groups in the US, Canada, and Australia, as well as in the UK (with, of course, the necessary ethical approval in place).

Personally, I’d nearly always suggest to avoid recruiting people you know, and particularly those you know well (and even more particularly your clients). There’s just too many opportunities for biases and demand characteristics to creep in. If they know your a priori assumptions, for instance, it may be very difficult for them to provide a contradictory view. Jasmine Childs-Fegredo adds:

You also need to consider what it may mean to recruit participants from a place you have worked.  For example, if you know the staff and some clients in an NHS service which you now want to recruit from, you will need to make sure you are aware of this ‘dual-role’ you have, and approach things with due sensitivity and considering the ethics around that.

Approaching prospective participants

Some general pointers when approaching people to participate in your research:

  • Be friendly. If you’re cold, disinterested, or aloof you’re likely to immediately put people off. And bear in mind that some times you won’t know that you’re coming across in that way, even if you don’t mean too. If, like me, you have a ‘resting bitch face’, or tend to write quite curt emails, then think about ways of conveying a warmer and more welcoming invite.

  • It’s nearly always better to personalise your approach: to individuals, to particular groups, to sectors of the population. Most of us hate getting very generic research invites that have obviously just gone out to hundreds of people. It’s immediately off-putting: What’s my incentive to take part if there are hundreds of others like me who can do the same? But if it’s a personalised email, tailored to me (e.g., ‘Dear Mick, given your position as a counselling psychology teacher….’) then I’m much more likely to respond. That, of course, makes the recruitment process more time consuming, but it’s generally worth the payoff: more prospective participants, and prospective participants who feel welcomed into the research, respected, and understood.

  • Communicate your passion and excitement for your work, and for learning from your prospective participants. If they see it means something to you, it’s more likely to feel meaningful to them too.

  • It generally a good idea to find ways in which prospective participants can express a very preliminary initial interest before making a more definitive commitment. (In psychology, this is known as the ‘foot in the door’ technique). So, for instance, you could invite people who may be interested to click on a hyperlink (for instance, to a Google Doc) where they can leave their email address to be contacted (make sure it is all GDPR compliant), or you could suggest that they email you for more information before making any commitment.

REcruitment materials

Recruitment materials are, essentially, the ‘adverts’ that you put out there to attract interest. And like any adverts, they have to be carefully thought through. A good place to start might be to reflect on the question (and discuss with peers), ‘What kind of research recruitment materials make me more likely to respond?’ For instance, is it where there’s a more personalised approach, or where it feels meaningful to your own life and concerns? And, importantly, also ask yourself, ‘What kind of research recruitment materials make me instantly hit “delete”?’ For instance, is it when you’re not clear what they’re asking you to do, or if it seems to go on and on with ever-finer details?

Some general pointers about recruitment materials:

  • Proof them, proof them, proof them… and then get a few friends/colleagues to proof them, proof them, proof them. If your prospective participants are anything like me, they’ll be really put off by misspelt emails or slapdash flyers which seems to change font half way through. After all, if you can’t put enough care into getting your spelling right, what’s that going to say about how you’ll treat your participants? As emphasised throughout this blog, trust is everything!

  • So try to get a balance between being friendly but professional. You don’t want to come across as too mechanistic or formal (it can feel intimidating), but too informal can also feel overly-casual and potentially unprofessional.

  • Be sensitive: for instance, putting a research request on Twitter for people who have been ‘traumatised’ could trigger all sorts of responses in some people—many of whom may never actually get in contact with you.

  • Don’t be pushy. You don’t want to put people under pressure to take part, or feel coerced in any way. So avoid headlines phrases like, ‘Please take part in my research,’ or ‘Participants needed,’ or ‘Would you like treatment for your anxiety?’

  • Having said that, help participants see the potential benefits of taking part: to themselves, to the therapeutic field, to their wider communities. Remember, chances are they’re looking for reasons not to take part rather than reasons to; so you need to consider what the incentives might be and make those explicit. Of course, as above, you don’t want to be pushy, and you also need to be explicit about any potential risks. But, for instance, many participants can find it really rewarding talking about their experiences, and this is a possibility you can highlight in your information sheet. Also, for many participants, there can be a great deal of value and meaning in contributing to the development of improved mental health treatments and services for all. So if that is a potential impact of your study (and you’ve got a coherent strategy for achieving it) you can make that clear in your recruitment materials.

  • Length of notices and other written materials is another challenge. Ethically, there’s a lot that you may want/have to say, but it can easily be overwhelming and off-putting for participants if it’s too much, too soon. One option is to disseminate, in the first place, just brief notices or flyers, that prospective participants can then follow up to find out more detailed information.

  • Tailor your materials to the specific audiences. For instance, a notice on the professional network LinkedIn might have a more serious tone than a post on Twitter, and a face-to-face invite may be framed in a very different way.

  • Be clear and concrete about what people should do next if they’re interested. Have your email address, for instance, in big and bold on the recruitment email, or a hyperlink for people to click on to sign up. Make it as easy as possible for potential participants to follow through. A good option here may be to have a website that people hyperlink to from social media platforms, that then has more information about the study and clear details of how to contact you. Generally, try to ensure that prospective participants can reach you through hyperlinking—if they have to copy your email/phone number down from a jpg, for instance, they may be a lot less likely to get in touch.

  • And finally, don’t be weird. It’s an obvious thing to say, and ‘weird’ can mean many different things to different people, but if a prospective participants wants to feel assured that it’s safe to take part, it’s best to keep quirkiness in how you approach people, and what you put on your recruitment materials, to a minimum.

people who know people

Understandably, people may be less likely to respond to your research request if they don’t know who you are. So if you know someone who has contact with prospective participants, you may want to ask them if they can help you in the recruitment process. A research or clinical supervisor, for instance, might have a wide network of people they’d be willing to forward an email on to, or to post on their social media sites. There may also be specialists in the field that you’re looking at who’d be willing to support you by forwarding on recruitment invitations. You can always ask. And you can also add some information about yourself on recruitment sites so that you are less anonymous: even a photo and a brief biography can help prospective participants feel that there is a real and friendly person behind the recruitment process.

If you are wanting to recruit clients into your study, one way of reaching them is through counsellors and psychotherapists. This has to be done with extreme sensitive, though, and without in any way breaching confidentiality. For instance, it would be entirely unethical to ask therapists to pass on contact details of their clients to you so you could email them directly! You also need to make sure that the therapists’ clients are not feeling under any pressure to participate: deference effects means that clients may feel obliged to say ‘yes’ to their therapists, even if they don’t want to. One workable option may be to ask psychotherapists and counsellors to pass on a flyer to their clients giving them information about your study, and then the clients can contact you, in their own time, if they are interested.

Making contact with prospective participants through professional, training, or service delivery organisations is another way of reducing the anonymity of your request and enhancing its ‘legitimacy’. Here, for instance, a counselling service might forward on a request from you to their counsellors or clients; or else they might make the request as an organisation themselves (with you identified as the researcher). In general, recruiting through an organisation can create quite a ‘containing’ frame for research, and in some cases—quite rightly—is the only way in which you would be able to access particular populations (for instance, service users of a domestic violence organisation). If you can align your research with the specific wants and needs of an organisation—for instance, if it will provide evidence on their service effectiveness—they may be particularly keen to support you in it.

Finally, on this point re anonymity, prospective participants may be much more likely to respond to you if they can get a sense of you, as a person, rather than as an unknown name on a flyer. So, for instance, if you can go along and do a talk—even 5 mins—at a service user group, or chat to people over a conference poster, that might really help with response rates. As Jasmine emphasised earlier, people need to feel that you’re safe to open up to: someone known and familiar rather than alien and strange.

Be responsive

If a prospective participant gets in touch with you, respond. Don’t leave it sitting in your email inbox for weeks. It’s an obvious thing to reiterate, but it’s essential to treat prospective participants with courtesy and respect.

If it’s not working…

If your recruitment strategies aren’t working, don’t panic! Give it a bit of time and see what emerges. But if, after a few weeks, you’re still not getting any eligible volunteers, it might make sense to start looking at what adjustments you might want to make.

First issue, of course, is where the ‘blockage’ might be. For instance, is it that no one is making initial contact with you about your research; or is it that they are, but then not following up when you reply. That should give you some clues about where adjustments may be required.

If no one is showing any interest, it generally makes sense—at least initially—to stick with your participant group and look at additional, or alternative, strategies for recruitment. Are there particular networks, for instance, you can make contact with; or alternative social media sites? Here, you may need to balance the coherence and homogeneity that comes from having participants from just one source, against the greater recruitment possibilities that come from broadening things out. With this issue, there’s no right answers; but one thing I would say is to try and have a few from each if you can. For instance, three participants recruited through Facebook and six from snowball sampling can be fine, and you might even be able to say something about the differences between them. But eight from Facebook and one from snowballing leaves the latter a bit of an ‘odd man out.’ We don’t know if their responses are specific to them or to the strategy they were recruited through.

If widening your recruitment strategies still isn’t working, you may need to revisit your participant group and, with it, the specific question you’re looking at. For instance, if you’re exploring the psychotherapy experiences of Kenyan men, would expanding that to East African men, or men across all of Africa, make for a more viable recruitment process? Here, as above, you’re striving to strike a balance between having a scope that is broad enough for successful recruitment, but narrow enough to make your research project meaningful and coherent. Again, no right answers; but being open to adjusting your design, where necessary, can be a real advantage.

Use supervision

Remember to make use of support from your research supervisor(s). Jasmine Childs-Fegredo says:

Should you be experiencing issues with recruitment, it’s worth getting in contact with your supervisor(s) in the first instance, to talk through what you could do going forwards, and then report back to them as and when things start moving or if you need further support. Supervisors generally have the experience to nip things in the bud early on, and may have ideas you have not previously thought of. It’s best not to just leave things, and expect things to get better without some support. Supervisors are busy people and may not be able to see you immediately, but it’s always worth getting an advance meeting in the diary with them to discuss where you are in your recruitment strategy.

What does the research say?

In thinking through strategies for recruitment, it may also be very helpful to consult the research on what works and doesn’t work, says our former PsychD Course Convenor, Mark Donati. For instance, you can find papers like, ‘Factors influencing recruitment to research: qualitative study of the experiences and perceptions of research teams’, or how about this one: ‘Swiss chocolate and free beverages to increase the motivation for scientific work amongst residents: a prospective interventional study in a non-academic teaching hospital in Switzerland’! There may also be papers on recruitment for your particular participant group, for instance, ‘Overcoming barriers to recruiting ethnic minorities to mental health research: a typology of recruitment strategies’ and ‘Recruitment and retention of older minorities in mental health services research’. When you write up your research project, being able to report that you used research, itself, to direct your methodological choices can look very impressive.

In conclusion

Plan, be realistic, be proactive, and flexibly adjust if things aren’t working out…. That sounds like the recipe for a successful life, so no surprises that it also holds for successful research recruitment. And, of course, as Rosie and Jasmine emphasised, be sensitive, collaborative, and kind. Even if that doesn’t get you the most participants, it’s the ethical and right thing to do. Remember that you’re part of a wider research community, and successful enquiry, across the board, requires research participants to feel like they are valued participants in that process—not just ‘subjects’ that get discarded when the research is done. So approach prospective participants with a spirit of genuine openness and dialogue.

Mark thought I should end this blog on an upbeat note, and he’s absolutely right. Yes, it’s hard work; yes, it can be a struggle; but the sense of satisfaction, excitement, and sheer relief you can get from having all your data finally collected—and in a robust, ethical, and caring way—is second to none. So, if it’s seeming (or feeling right now), like an uphill struggle, keep your eye on that prize. With proactiveness, persistence, and creativity, you’ll get there for sure.


Acknowledgements

Thanks to trainees and tutors on the PsychD Counselling Psychology Programme at the University of Roehampton for suggestions and advice. Photo: Maya, by Daniel Walford.

DISCLAIMER

The information, materials, opinions or other content (collectively Content) contained in this blog have been prepared for general information purposes. Whilst I’ve endeavoured to ensure the Content is current and accurate, the Content in this blog is not intended to constitute professional advice and should not be relied on or treated as a substitute for specific advice relevant to particular circumstances. That means that I am not responsible for, nor will be liable for any losses incurred as a result of anyone relying on the Content contained in this blog, on this website, or any external internet sites referenced in or linked in this blog.

Applying for a PhD in Counselling and Psychotherapy: Some Pointers

I’m sometimes asked about the process of applying for a PhD in counselling and psychotherapy and whether it’s worth doing, so I wanted to put together some pointers. Just to say, this blog is written from a personal perspective, for study within a UK context, and the focus is on research-based PhDs rather than professional doctorates. More on that distinction below.

Why should I want to do a PhD?

There’s probably a good chance that you shouldn’t. Yes, it’s pretty cool being able to write ‘Dr’ before your name when you fill in forms (at least for the first few times) but a PhD is nearly always a long, hard slog of 3 to 4 years or more (mine took about eight!): moments of insight, excitement, and achievement interspersed with long periods of boredom, frustration, and sheer hard work. Then there’s the emotional toil; and, like your original counselling or psychotherapy training, it can play havoc with your relationships. So don’t ask yourself whether you want to do a PhD when you’re feeling inspired, eager, and motivated. Ask yourself after a long, hard day’s work when all you want to do is pour yourself a glass of wine and flop down, mindlessly, in front of The Great British Bake Off. And, if you are going to try and do a PhD, make sure you really know why. I’d say that doing a PhD makes sense if you:

  • Want to go into academia/teaching as a career.

  • Want to go into research as a career (though there are very limited options here).

  • Really, really love research and want to spend a long period of time immersed in it.

  • Have a specific area of interest that you are really committed to making an original and significant contribution to.

I’m sure there are other reasons, but they need to be really good ones, and ones that are going to sustain you over the course of the programme. If your reason for applying is just that you’re not really sure what to do next, there’s a good chance that the hard slog of a PhD is not for you.

So what actually is a PhD in Counselling/Psychotherapy?

A PhD is generally a series of research studies, culminating in the writing of a dissertation (or ‘thesis’) of 80,000 words or so. Essentially, you’re writing a book, but one based on some systematic research process. Before you do anything else, have a look at some counselling or psychotherapy PhD theses to get a feel for what you’ll need to do: for instance, Adam Gibson’s Shared decision-making in counselling and psychotherapy (2019, University of Roehampton) or Katie McArthur’s Effectiveness, process and outcomes in school-based humanistic counselling (2013, University of Strathclyde).

A PhD programme doesn’t generally have a clinical component, and there’s often only a small amount of structured teaching—usually around research methods. Generally, the bulk of the work is self-study, alongside regular meetings with your supervisor(s) (perhaps 1-2 hours, once a month or so). PhDs can usually be undertaken on a full time basis (taking around 3 to 5 years) or part-time basis (4 to 6 years, and sometimes more).

A ‘PhD’ is generally a wholly research-based program of study, and is different from a ‘professional doctorate’, which tends to have a more clinical, professional, and/or reflective element (see, for example, the Metanoia Institute Doctorate in Psychotherapy by Professional Studies, or the University of Chester Doctor of Professional Studies in Counselling and Psychotherapy). These latter courses offer a more holistic programme of development for qualified counsellors or psychotherapists and often make more sense to undertake—unless your interest is solely on the research side of things.

A PhD is also very different from a doctorate in counselling psychology or clinical psychology, like our PsychD in Counselling Psychology at the University of Roehampton. These courses are for graduates in psychology and offer a full professional training from start to finish.

What should i focus on?

Generally, it’s good to start the process of exploring PhDs with some idea of what you want to look at (pointers on choosing a research topic can be found here). This doesn’t need to be fully formed—indeed, it’s important that you’re open to input from prospective supervisors—but having some sense of the field that you want to look at, the kinds of questions that you want to ask (and, perhaps, the method you might adopt) is important in being able to take things forward. So, for instance, you might want to look at something like, ‘Autistic children’s experiences of counselling,’ or ‘The role of empathy in psychotherapy with older adults,’ or ‘A phenomenological analysis of transference.’ Ideally, it’s good to write this up as maybe a page or so of ideas, so that it’s something you can send out to prospective supervisors to start a discussion about your ideas.

Should I approach potential supervisors?

Yes. You don’t have to, but I would generally suggest you find the leading academics in your subject area, or the particular method you’re wanting to adopt, and email them to find out if it’s something that they might, potentially, be interested in supervising. When you do that, it’s important to have some idea of what it is that you want to do; and the brief, one page sketch, as detailed above, is the kind of thing you can send them to let them know more. That’s the kind of thing that works for me if someone approaches me in this regard: if it’s very vague and open (‘I’m thinking of doing a PhD, sort of, maybe, what do you think?’) it can be a bit frustrating; if someone sends through screeds and screeds of an extremely detailed proposal, it can feel a bit overwhelming and like there’s not much flexibility there (but better the latter than the former).

Bear in mind that, generally, academics will only take on a small number of PhD students, so for them to want to work with you it has to be very much in their subject area. For instance, I’d be interested in PhD proposals on subjects like relational depth, or humanistic counselling in schools, or existential therapy; but if someone approached me with a PhD proposal for Transactional Analysis, even if I might think it was a great idea, I wouldn’t feel able to take it on. If you approach someone, though, you can always ask them to let you know other potential supervisors who might be more appropriate.

Can I Apply Directly to a University?

Yes, you can do it that way too. For instance, you could directly apply to the University of Roehampton here. (In fact, even if you have spoken to an academic who’s expressed interest in working with you, you would still need to formally apply through such channels.) If you’ve got a strong PhD application a university will probably give it close consideration whether or not they’ve got a specialist in the specific area. However, the advantage of approaching an academic first is it gives you some time to refine your proposal in line with what they may see as the key, or best, questions in that area. Often, there’s an iterative process of some initial informal discussion with an academic, maybe a refining of the research question, then a formal application—after which, of course, there’s further refinement and development of the research plan.

Where should I apply to?

There’s lot of different universities where you can do PhDs on counselling and psychotherapy topics. Sometimes that will be in a department of psychology, sometimes within a particular counselling or psychotherapy unit. Sometimes as part of an educational degree. In theory, pretty much any university should allow you to apply there for a PhD in the counselling and psychotherapy field.

Given that research meetings often aren’t that frequent, and can often be conducted online, geographical proximity needn’t be a major consideration. For instance, I’ve worked well with PhD students at the other end of the UK, as well as in mainland Europe. PhD programmes that have some taught elements will require some face-to-face attendance though. Also, at least a little face to face meetings with a supervisor—even if it’s only once a year or so—is generally a good idea (excepting COVID-19!).

So I’d tend to say apply to a university based on where the best supervisor(s) is going to be. That is, someone who knows the areas (or methods) you’re interested in and has published in it, has shown interest and motivation if you’ve approached them, and feels able to support you in your research programme. One thing you really don’t want is to end up with a supervisor allocated to you who feels that they’re having to take you on. That’s rare: but being proactive in identifying the right supervisor, liaising with them, and then applying to the respective university is generally the best way of ensuring you’ll get the support you need.

Also, there may be advantages in applying to a university which has a group of students doing PhDs in related areas, so that you have a community around you to discuss your work with, learn from others, and get support. That’s something you can find out from the academics there, or ask on interview. If the university has an active culture of psychotherapy and counselling research, that’s also probable good sign. Do they have a research centre in this area, for instance (like our CREST Research Centre at the University of Roehampton), or seminars, or do academics and students from this university regularly attend conferences like the annual BACP Research Conference? Having that active, engaged community around you may be really important in sustaining your interest and motivation over the course of the programme. You really don’t want to do this all on your own.

What qualifications do I need?

In most instances, the main thing to show is that you have experience of research, ideally in the counselling and psychotherapy field. So a Master’s in the area (for instance, an MSc in Research Methods) would be ideal, or a Master’s in counselling or psychotherapy which involved some significant research component. If you don’t have that, then experience of research in the workplace could count: for instance, if you have been working for several years in an evaluation capacity. Demonstrating motivation and interest in research, as well as a viable research proposal, is also very important. For the institution and supervisors, taking on a PhD student is a big commitment, so they really need to feel that you will be in it for the full long haul.

who’s going to pay me to do it?

Probably yourself. Unfortunately, there’s very little funding available for PhDs in the counselling and psychotherapy field, and most students do pay for it themselves. There are some exceptions—for instance, universities may have scholarships that they award on an intermittent basis, and there are grant funding bodies like the ESRC—but it’s generally extremely competitive and if you go down these routes you may have to do your PhD about a particular topic that the institution is interested in.

What happens once I’ve applied?

The academics at the university you’ve applied to will consider your application, in the light of the kinds of criteria discussed above, and may well invite you for interview to discuss your application further. If you’re accepted, you can then get going with refining your research project, and preparing to run your study.

In conclusion

I wouldn’t want to put anyone off applying for a PhD in counselling or psychotherapy. It’s got the potential to be an amazing journey: with discovery, in depth engagement with your topic, and the opportunity to make a unique contribution to the counselling and psychotherapy field. Relationally, too, it can be a unique opportunity to engage with peers, academics, and participants. You become a world-leading expert in your field; and if you want to go into academia or a research job, it’s pretty much essential. But it is a massive commitment, and you really need to be realistic about what you are letting yourself in for before you embark on it. As a PhD student recently said to me:

The ideal position to do a PhD is one where you know the route is hard, less than ideal, uncertain, but it is also the necessary route.

Very best of luck with it.


Further Reading

Hayton, J. (2015) PhD: An uncommon guide to research, writing and PhD life. James Hayton PhD. Suggested by a PhD student as very realistic and enjoyable.

Acknowledgement

Thanks for the guidance from current and former PhD students on the content here.


Disclaimer

The information, materials, opinions or other content (collectively Content) contained in this blog have been prepared for general information purposes. Whilst I’ve endeavoured to ensure the Content is current and accurate, the Content in this blog is not intended to constitute professional advice and should not be relied on or treated as a substitute for specific advice relevant to particular circumstances. That means that I am not responsible for, nor will be liable for any losses incurred as a result of anyone relying on the Content contained in this blog, on this website, or any external internet sites referenced in or linked in this blog.

Just to add, no liability will attach to the University of Roehampton as a result of the training and consultancy work presented on this website, which I am carrying out in a private capacity.

Essay Writing in Counselling and Psychotherapy: Top Tips

I’m a liberal when it comes to most things—except (as my students will know) fonts, formatting, and grammar. So why am I a fully signed-up member of the Grammar Police (or should that be ‘grammar police’)? Well, aside from my various OCDs (yup, that’s Oxford Comma Disorder), it’s a way that you, as a writer, can make sure that your beautiful, brilliant, creative writing is seen in its best possible light—not detracted by missing apostrophes and torturously convoluted sentences. So here are over 25 top tips for those of you writing essays and dissertations—at all levels—based on years of marking and encountering the same issues time after time. All of these tips are aligned with the Publication Manual of the American Psychological Association (7th edition), which provides an essential set of guidelines and standards for writing papers in psychology-related fields. There’s also a checklist you can download from here to go through your draft assignments to check everything is covered. (And just to say, by way of disclaimer, listen to your tutors first and foremost: if they see things differently, do what they say—they’re going to be the ones marking your papers!)

Puncutation

  • Apostrophes. You just would not believe how many students working at graduate, Master's, and even doctoral level dont know when to put apostrophe's and when not to. Check out the rules on it—it takes two minute's on the web (try this site)—and you'll never drive your marker's crazy again (whose this Roger’s bloke that students keep writing about?).

  • Single (‘ ’) or double (“ ”) quotation marks? For UK English it’s single; for US English it’s double. The only exception is when you give quotation marks within quotations marks, in which case you use the other type. So, for instance, in UK English you might write:

    • Charlie said, ‘I’ve often told myself, “buck up, don’t be stupid,” but I do find it hard.’ On the other hand, Sharon said…

    And while we’re at it, make sure those are ‘curly marks’ (or ‘smart apostrophes’), and not the symbols for inches (") or feet ('), which are straight.

  • One space after a full stop. Not two. That’s for when we had typewriters.

  • Colon (:) before a list, not semi-colon (;), and definitely not colon-dash (:-).

  • Write out numbers as words if they are below 10 (except if they are to do with dates, times, or mathematical functions; or at the start of the sentence). So, for instance:

    • ‘Across the three cohorts there were over 500 participants.’

    • ‘In this study, six of the young people said…’

  • Think where you’re putting your commas. They’re not sprinkles: something you just liberally and randomly scatter over your text. So check where you’ve put them, and that they meaningfully separate out clauses, or items, in your writing.

  • And, while we’re at it, a comma before the last item in a list (after ‘and’). This is known as an ‘Oxford comma’, and is recommended by the American Psychological Association (APA) to improve clarity. So, for instance, you’d write that ‘Across the counselling, psychotherapy, and psychiatric literature…’ rather than ‘Across the counselling, psychotherapy and psychiatric literature…’

  • Watch out for over-capitalising words. In most cases, you don’t need to capitalise—you’re not writing German (unless, of course, you are). Most words don’t need capitalisation (e.g., ‘person-centred therapy’, not ‘Person-Centred Therapy’), unless they are ‘proper nouns’ (that is, names of specific one-of-a-kind items, like Fritz Perls or the University of Sussex).

  • Key terms should be italicised on first use. Say you’re writing an essay about phenomenology, or it’s a key term that you’re going to define subsequently. The first time you use the term, italicise it. For instance, ‘Person-centred therapy is based upon a phenomenological understanding of human being. Phenomenology was a philosophy developed by Husserl, and refers to…’. An exception to this is that, if you want to introduce a term but without any subsequent definition (perhaps it’s not that central to your essay), put it in quotation marks. For instance, ‘Transactional analysis is based on such concepts as “ego states” and “scripts”, while Gestalt therapy…’

Quotations AND CITATIONS

  • Reference your claims. Whenever you state how things are, or how things might be seen, reference where this is from. Typically, a paragraph might have four or more references in it. If you find that you have several paragraphs without any at all, check you’re not making claims without saying their source. If it’s your own opinion, that’s fine (particularly later on in essays, for instance in the discussion), but be clear that that’s the case.

  • If you give a direct quotation, give the page number of the text it’s from (as well as the author(s) and date).

  • If a quotation is more than 40 words (a ‘block quotation’), indent it.

  • Otherwise, treat direct quotations as you would other text. So you don’t need to italicise it, put it in font size 8 or 18, use a different font colour etc. The same for quotations from research participants: use quotation marks and treat as block quotations if over 40 words, but otherwise leave well enough alone.

  • The page number comes between the close quotation mark and the full stop (if the direct quote is in the text). For instance: Rogers (1957) said, ‘The greatest regret in my career is that I didn’t develop pluralistic thinking and practice’ (p. 23). The only exception to this is with block quotations, in which case the page number comes after the full stop. Stupid, I know, but there you go.

  • In text citations for papers with 3 or more authors only need the first author now from first citation onwards, with ‘et al.’: e.g., ‘Cooper et al. (2021) say…’

Paragraphs, Sentences, and Sections

  • One paragraph, one point. Don’t try and squeeze lots of different points and issues into one paragraph. Often, a good way to write paragraphs is with a first sentences that summarises what you are saying in it, then subsequent sentences that unpack it in more detail.

  • Keep sentences short. In most cases, it doesn’t need to be more than three lines or so. If it’s longer, check whether you can break the sentence down into simpler parts.

  • Keep sentences simple. You don’t normally need more than two or three ‘clauses’; and if you’ve got more, for instance, like this sentence has—with lots of commas, semi-colons, and dashes in it—you can see how it starts to get more difficult to follow, so try and simplify.

  • Make sure you give clear breaks between paragraphs. So that the reader can see where one ends and the other begins. For instance, have a line break, or else indent the first line of each paragraph.

  • Headings should stand out. That’s what they are there for, so make sure they are different from the rest of the text. For instance, do as bold and centred. Also, if you are using different levels of headings (for instance, headings, subheadings, and sub-subheadings), make it really clear which are which, with higher levels more prominent in the text.

  • Don’t forget page numbers. If you want your assessors to be able to give feedback, they need to be able to point to where things are.

General writing

  • Use acronyms sparingly. ‘The AG group felt that ACT was superior to CBB on the TF outcomes…’ Unless you’ve got the memory of a child genius they’re a nightmare. If you do use them, make sure you explain what they are on first use.

  • Avoid jargon/overly-casual terms. ‘The therapists in the study seemed quite chilled; but, for future research more groundedness and heart-centredness could possibly help.’ Enough said!

  • Avoid repetition. Saying something once is nearly always enough. You don’t need to repeat it again and again. It gets tedious. Especially when you say things over and over again.

  • Be consistent in the terminology that you use. For instance, if you are doing an interview study with young people, don’t switch randomly between calling your participants ‘young people’, ‘adolescents’, ‘teenagers’, ‘clients’, and ‘participants’. Choose one term and stick to it and; if you do use more than one term, be consistent in which one you use when.

  • Use footnotes/endnotes sparingly. It can be frustrating for a reader to jump between your main text and then subtexts written elsewhere. So try and include everything in your main text if you can (for instance in parenthesis).

  • Don’t assume your readers know what things mean. ‘When it comes to measures based on normative, formative indicators…’ What? You don’t know what ‘normative’ and ‘formative’ mean (and it’s not a music group, though the name ‘The Normative Formatives’ is pretty cool!). The point here, as above, is to spell things out so that the reader knows what you are talking about. If it’s brief you could do that in parenthesis in the sentence. If not, give it dedicated sentences.

  • Check the spell and grammar checkers. Those wiggly blue and red lines underneath your writing (on Microsoft Word) do mean something. Sometimes it’s just the software being over-sensitive, but it’s always worth checking and seeing what it’s picking up. If you’re software doesn’t do spell and grammar checks, it might be time to upgrade. You need something or someone else to give this a thorough check through before submitting any piece of work.

  • Make your file names meaningful. And finally, if you are sending out documents for assessments as digital files, give it a name that is going to mean something in someone else’s system. ‘Essay.doc’ or ‘Berne version 3 final’ is really not going to help your assessor know which is your submission—particularly in the midst of tens or hundreds of others. So make sure your surname is in the file title (unless the submission needs to be anonymised), and add a reference to the specific assignment: for instance, ‘Patel case study 1’. Adding a date of submission, or completion, is also very useful, though I would suggest always doing this in the format ‘year-month-day’ (rather than ‘day month year’), so that computers store more than one version of the file in the correct order (assuming the files are sorted alphabetically). So that gives you a file title like ‘Patel case study 1 2020-03-10’ and, with a name like that, it’s unlikely to get mixed up with anything else.


With many of these ‘rules’, the main thing is to be consistent. For instance, most markers won’t mind if you use double quotation marks rather than single, or italicising all your quotes, but the key thing is to do it all the way through. It’s when it’s changing that it gets confusing, because the reader thinks you might mean something by it, when in fact it just means you weren’t thinking about it. But how do we know?! Bear in mind, in particular, that your marker may have several assignments to work through, so anything that can help make their life easier is likely to be worth it. And the great thing is, once you get into these habits, they’ll stick with you for next time. As your academic level progresses, there will be more and more expectation that you’ll get these things ‘right’. So use the checklist to go through your first few assignments, and also ask a peer to scrutinise it using the checklist, and once you’re finding that you’re addressing the issues from the start you can stop using it.

Last thing, and I’ve already said this (so much for avoiding repetition!), but for a brilliantly concise and comprehensive guide to academic writing, go to the Publication Manual of the American Psychological Association (now in its seventh edition). Keep it by your writing desk, your bedside, your toilet…. it’s an invaluable investment in terms of getting through your assignments, because it gives you a consistent and clear set of guidelines on everything from referencing to headings to writing style.

Actually, sorry, really really last thing, and I couldn’t end this blog without saying it because my students won’t recognise me. Times New Roman 12 point. That’s all you need. No Comic Sans, no Bahnschrift Light SemiCondensed. Just one, nice, clear font all the way through.

Keep it simple and let the glorious light of your creative genius through. Good luck!

Acknowledgements

Photo by Lovefreund

The Introduction: Some Pointers

The following blog is for Master’s or doctoral level students writing research dissertations in the psychological therapies fields. The pointers are only recommendations—different trainers, supervisors, and examiners may see things very differently.

What does an Introduction do?

The aim of an introductory section is to help your reader understand what your dissertation is about and why it is important. It is also an opportunity to help them understand the context for your study so that they can understand where it is coming from and what it is trying to contribute to the wider field. 

An Introduction will typically include the following sections, though not necessarily in this order:

  • Aims/objectives of the research

  • Research question(s)

  • Personal rationale

  • Contextual rationale

  • Background literature

  • Definitions of key terms

  • Outline of the dissertation

For a dissertation, an Introduction is often separate from a Literature Review. The former is often the place where you set out why you are asking this question(s), whereas the latter sets out what you already know in answer to this question(s).

Aims/Questions

What is the purpose of your research? Your Introduction is the place to try and state, as explicitly as possible, what your research aims and/or questions are (see pointers here).

Personal Rationale

So why are you doing it? Why is it important to you? In most therapeutic fields, it is entirely legitimate (if not essential) to say something of why you are coming to this question, at this point in time. And the deeper you can go into your own personal rationale, the more insightful and authentic your personal account is likely to be. So some questions you might want to ask yourself are:

  • Why this research question/topic area?

  • Why does it matter to you?

  • What does it mean to you?

  • Why now?

  • What was your personal journey towards this research question?

  • How do you feel about this research question? What emotions are generated in you when you think about it?

  • How does this research question connect to your:

    • life

    • personal history

    • identity

    • values and meanings

    • aspirations for the future?

Something you might find really helpful is to do this as an exercise with a partner. Ask them to interview you, say for 20 minutes, using these questions. Record it and then listen back once the interview is over. That can really free you up to talk honestly and openly about some of the concerns and motives that underpin why you are doing this work. And, of course, you don’t need to share it all in your Introduction: but knowing where you want to go and why is a critical part of conducting an informed, in-depth, and self-reflexive study.

As part of this reflexive work, you might also want to ask yourself the question, ‘Are there some particular answers that I, consciously or unconsciously, would “like” to find?’ When it comes to writing about your personal biases in relation to the research question, however, that may be more likely to go in your Methods section. Here, in the Introduction, the focus is more on biases and assumptions that may have led you to ask this question in the first places.

Contextual Rationale

Of course, it’s not all about you. There’s also got to be good reasons, for the wider field, in you asking these questions at this point in time. For instance, maybe there’s a lot of research on how young people experience acceptance and congruence, but not empathy; or perhaps there’s evidence of particular increases in mental health problems in young people of Asian origin, so we really need to know what can help them.

So your Introduction is also a place where you can say about why your research is of importance in the grander scheme of things. Use evidence wherever you can, though it might be historical or socio-political as well as psychological.

Again, it can be really helpful to explore these questions in a pair. Get interviewed by a colleague, but this time invite them to probe you on why they should care about what you are doing. Some questions that they might want to ask/role-play are:

  • Why should I (as a counsellor/psychotherapist/counselling psychologist/researcher/commissioner/policy-maker) care about what you are doing?

  • What is it going to teach me, as a counsellor/psychotherapist/counselling psychologist/researcher/commissioner/policy-maker?

  • Don’t people already know the answer to your question? How is it going to add to the literature out there?

  • Why is it worth anyone spending time on this?

  • How will it make a contribution to:

    • Society?

    • Clients?

    • Other therapists?

    • The people who took part in the study?

Have you convinced them it is worthwhile (indeed, have you convinced yourself)? If not, it may be worth spending some time thinking through what it is that you really want to do, and whether it really is important. It may be that you sense it, it’s just difficult putting it into words. But try and find that sense so that you have a really clear basis to underpin your research work.

Background literature

Your Introduction is also a good place to explain anything that the reader needs to know about to understand the context and meaning of your study. For instance, how many young people enter person-centred therapy every year, or how did the concept of ‘alliance ruptures’ emerge and what are it’s theoretical underpinnings.

Of course, you’re also going to be reviewing the background literature in your Literature Review chapter, so how do you know what goes where? Maybe the best way to think about this, as above, is that content for the Literature Review chapter provides preliminary answers to your questions, whereas content for the Introductory chapter helps you understand what the question is and why it’s important. So, for instance, in our study of young people’s experiences of empathy, literature on how Rogers defined empathy might go in our Introduction, as might literature on mental health problems in adolescents. But findings of, for instance, a quantitative study on how young people rated the importance of empathy would go in our Literature Review, because it’s providing us with some important initial answers to the question we are asking.

Defining Key Terms

Closely related to this, what we can also do in our Introduction is to define key terms: anything that the reader is going to need to understand to be able to make sense of our thesis; and also so that they know how we, specifically, are choosing to use certain terms. For instance, do we mean ‘empathy’ as Rogers defined it, or as neuroscientists have understood it, or in the Kohutian sense? That’s very important information for the reader in terms of understanding our work.

What about if we want to leave the definition(s) open to our participants rather than imposing on them a particular understanding? Indeed, maybe our research is about exploring what young people understand by empathy, or what alliance ruptures mean to clients.

Research questions of this type (‘What do people understand by x?’) can be great, particularly if we’re coming to our research from a very inductive, ‘grounded’ epistemological position. However, I would say that it is a case of either/or: that is, either ask about what something means, or ask about how it is experienced/what it does—but don’t try asking both of these questions at the same time. Otherwise, you’re essentially asking your participants to describe the experience/effects of lots of different things, and you’re not likely to come up with a particular coherent answer. If Person A, for instance, defines empathy as Z, and experiences it as V; and Person B defines empathy as Y, and experiences it as U; then we may have learnt about different definitions of empathy, but our findings of V and U don’t really mean much because they refer to different things (Z and Y).

Outline Structure

Finally, your Introduction is a place where you can say what your thesis is going to look like: leading the reader through the different chapters of your work so that they know what is to come. You don’t need to do too much detail, maybe just a page or so, but something that gives them a clear and coherent sense of the route ahead.

Conclusion

By the end of your Introduction, your reader should have all they need to embark on the journey of your thesis; and, ideally, be motivated and excited to travel forward. So do make sure, as you describe your reasons for doing this work, or what the work is about, that you also draw the reader in: interest them, compel them, make them want to know more. Think of it like a tourist guide preparing your traveller for a trip ahead. Tell them what they need to know, but also not everything. After all, you want them to experience it first hand, and to learn what you have learnt as you travelled into the heart of your research.

DISCLAIMER

The information, materials, opinions or other content (collectively Content) contained in this blog have been prepared for general information purposes. Whilst I’ve endeavoured to ensure the Content is current and accurate, the Content in this blog is not intended to constitute professional advice and should not be relied on or treated as a substitute for specific advice relevant to particular circumstances. That means that I am not responsible for, nor will be liable for any losses incurred as a result of anyone relying on the Content contained in this blog, on this website, or any external internet sites referenced in or linked in this blog.

Evaluating and Auditing Counselling and Psychotherapy Services: Some Pointers

How do you go about setting up an evaluation or audit of your therapy service—whether it’s a large volunteer organisation or your own private practice?

Clarifying your Aims

There’s lots of reasons for setting up a service evaluation or audit, and being clear about what your’s are is a vital first step forward. Some possible aims might be:

  • Showing the external world (e.g., commissioners, policy makers, potential clients) that your therapy is effective.

  • Knowing for yourself, at the practitioner or service level, what’s working well and what isn’t.

  • Enhancing outcomes by providing therapists, and clients, with ‘systematic feedback’.

  • Developing evidence for particular forms of therapy (e.g., person-centred therapy) or therapeutic processes (e.g., the alliance).

And, of course, there’s also:

  • Because you have to!

Choosing an Evaluation Design

There’s lots of different designs you can adopt for your evaluation and audit study, and these can be combined in a range of ways.

Audit only

This is the most basic type of design, where you’re just focusing on who’s coming in to use your service and the type of service you are providing.

Pre-/Post-

This is probably the most common type of evaluation design, particularly if your main concern is to show outcomes. Here, clients’ levels of psychological problems are assessed at the beginning and end of therapy, so that you can assess the amount of change associated with what you’re doing.

Qualitative

You could also choose to do interviews with clients at the end of therapy about how they experienced the service. A simpler form of this would be to use a questionnaire at the end of treatment. John McLeod has produced a very useful review of qualitative tools for evaluation and routine outcome monitoring (see here).

Experimental

If you’ve got a lot of time and resources to hand—and/or if you need to provide the very highest level of evidence for your therapy—you could also choose to adopt an experimental design. Here, you’re comparing changes in people who have your therapy with those who don’t (a ‘control group’). These kinds of studies are much, much more complex and expensive than the other types, but they are the only one that can really show that the therapy, itself, is causing the changes you’ve identified (pre-/post- evaluations can only ever show that your therapy is associated with change).

Choosing Instruments

There’s thousands of tools and measures out there that can be used for evaluation purposes, so where do you start?

Tools for use in counselling and psychotherapy evaluation and audit studies can be divided into three types. These are described below and, for each type, I have suggested some tools for a ‘typical’ service evaluation in the UK. Unless otherwise stated, all these measures are free to use, well-validated (which means that they show what they’re meant to show), and fairly well-respected by people in the field. All the measures described below are also ‘self-rated’. This means that clients, themselves, fill them in. There are also many therapist- and observer-rated measures out there, but the trend is towards using self-rated measures and trusting that clients, themselves, know their own states of mind best.

Just to add: however tempting it might be, I’d almost always you not to develop your own instruments and measures. You’d be amazed how long it takes to create a validated measure (we once took about six years to develop one with six items!) and, if you create your own, you can never compare your findings with those of other services. Also, for the same reason, it is almost always unhelpful to modify measures that are out in the public domain—even minimally. Just changing the wording on an item from ‘often’ to ‘frequently’, for instance, may make a large difference in how people respond to it.

Outcome Tools

Outcome tools are instruments that can be used to assess how well clients are getting on in their lives, in terms of symptoms, problems, and/or wellbeing. These are the kinds of tools that can then be used in pre-/post-, or experimental, designs to see how clients change over the course of therapy. These tools primarily consist of forms with around 10 ‘items’ or so, like, ‘I’ve been worrying’ or ‘'I’ve been finding it hard to sleep’. The client indicates how frequently or how much they have been experiencing this, and then their responses can be totalled up to give an overall indication of their mental and emotional state.

Its generally good practice to integrate clients’ responses to the outcome tools into the session, rather than divorcing them from the therapeutic process. For instance, a therapist might say, ‘I can see on the form that this has been a difficult week for you,’ or, ‘Your levels of anxiety seem to be going down again.’ This is particularly important if the aim of the evaluation is to enhance outcomes through systematic feedback.

General

A popular measure of general psychological distress (both with therapists and clients), particularly in the UK, is:

This can be used in a wide range of services to look at how overall levels of distress, wellbeing, and functioning change over time. A shortened, and more easily usable version of this (particularly for weekly outcome monitoring, see below), is:

Another very popular, and particularly brief, general measure of how clients are doing is:

Two other very widely used measures of distress in the UK are:

The PHQ-9 is a depression-specific measure, and the GAD-7 is a generalised-anxiety specific measure, but because these problems are so common they are often used as general measures for assessing how clients are doing, irrespective of their specific diagnosis. They do also have the dual function of being able to show whether or not clients are in the ‘clinical range’ for these problems, and at what level of severity.

Problem-specific

There are also many measures that are specific to particular problems. For instance, for clients who have experienced trauma there is:

And for eating problems there is:

If you are working in a clinic with a particular population, it may well be appropriate to use both a general measure, and one that is more specific to that client group.

Wellbeing

For those of us from a more humanistic, or positive psychology, background, there may be a desire to assess ‘wellness’ and positive functioning instead of (or as well as) distress. Aside from the ORS, probably the most commonly used wellbeing measure is:

There’s both a 14-item version, and shortened 7-item version for more regular measurement.

Personalised measures

All the measures above are nomothetic, meaning that they have the same items for each individual. This is very helpful if you want to compare outcomes across individuals, or across services, and to use standardised benchmarks. However, some people feel that it is more appropriate to use measures that are tailored to the specific individual, with items that reflect their unique goals or problems. In the UK, probably the best known measure here is:

This can be used with children and young people as well as adults, and invites them to state their specific problem(s) and how intense they are. Another personalised, problem-based tool is:

If you are more interested in focusing on clients’ goals, rather than their problems, then you can use:

Service Satisfaction

At the end of therapy, clients can be asked about how satisfied they were with the service. There isn’t any one generic standard measure here, but the one that seems to be used throughout IAPT is:

Children and young people

The range of measures for young people is almost as good as it is for adults, although once you get below 11 years old or so the tools are primarily parent/carer- or teacher-report. Some of the most commonly used ones are:

  • YP-CORE: Generic, brief distress outcome measure

  • SDQ: Generic distress outcome measure, very well validated and in lots of languages

  • CORS: Generic, ultra-brief measure of wellbeing (available via license)

  • RCADS: Diagnosis-based outcome measure

  • GBO Tool: Personalised goal-based outcome measure

  • ESQ: Service satisfaction measure.

A brilliant resource for all things related to evaluating therapy with children and young people is corc.uk.net/

Process Tools

Process measures are tools that can help assess how clients are experiencing the therapeutic work, itself: so whether they like/don’t like it, how they feel about their therapist, and what they might want differently in the therapeutic work. These are less widely used than outcome measures, and are more suited to evaluations where the focus is on improving outcomes through systematic feedback, rather than on demonstrating what the outcomes are.

Probably the most widely used process measure in everyday counselling and psychotherapy is:

  • SRS (available via license)

This form, the Session Rating Scale, is part of the PCOMS family of measures (along with the ORS), and is an ultrabrief tool that clients can complete at the end of each session to rate such in-session experiences as whether they feel heard and understood.

For a more in-depth assessment of particular sessions, there is:

This has been widely used in a research context, and includes qualitative (word-based) as well as quantitative (number-based) items.

Several well-validated research measures also exist to assess various elements of the therapeutic relationship. These aren’t so widely used in everyday service evaluations, but may be helpful if there is a research component to the evaluation, or if there is an interest in a particular therapeutic process. The most common of these is:

This comes in various version, and assesses the clients’ (or therapists’) view of the level of collaboration between members of the therapeutic dyad. Another relational measure, specific to the amount of relational depth, is:

A process tool that we have been developing to help elicit, and stimulate dialogue on, clients’ preferences for therapy is:

This invites clients to indicate how they would like therapy to be on a range of dimensions, such that the practitioner can identify any strong preferences that the client has. This can either be used at assessment, or in the ongoing therapeutic work. An online tool for this measure can be accessed here.

Interviews

If you really want to find out how clients have experienced your service, there’s nothing better you can do than actually talk to them. Of course, you shouldn’t interview your own clients (there would be far too much pressure on them to present a positive appraisal) but an independent colleague or researcher can ask some key questions (for instance, ‘What did you find helpful? What did you find unhelpful? What would you have liked more/less of?) which can be shared with the therapist or the service more widely (with the client’s permission). There’s also an excellent, standardised protocol that can be used for this purposes:

Note, as an interviewing approach has the potential to feel quite invasive to clients (though also, potentially, very rewarding), it’s important to have appropriate ethical scrutiny here of procedures before carrying these out.

Children and young people

Process tools for children and young people are even more infrequent, but there is the child version of the Session Rating Scale:

Demographic/Service Audit Tools

As well as knowing how well clients are doing, in and out of therapy, it can also be important to know who they are—particularly for auditing purposes. Demographic forms gather data about basic characteristics, such as age and gender, and also the kinds of problems or complexity factors that clients are presenting with. These tools do tend to be less standardised than outcome or process measures, and it’s not so problematic here to develop your own forms.

For adults, a good basic assessment form is:

For children and young people, one of the most common, and thorough, forms is:

Choosing Measurement Points

So when are you actually going to ask clients, and/or therapist, to complete these measures? The demographic/audit measures can generally be done just once at the beginning of therapy, although you may want to update them as you go along. Service satisfaction measures and interviews tend to be done just at the end of the treatment.

For the other outcome and process measures, the current trend is to do them every session. Yup, every session. Therapists often worry about that—indeed, they often worry about using measures altogether—but generally the research shows that clients are OK with it, provided that they don’t take up too much of the session (say not more than 5-10 minutes in total). So, for session-by-session outcome monitoring, make sure you use just one or two of the briefer forms, like the CORE-10 or SRS, rather than longer and more complex measures.

Why every session? The reason is that clients, unfortunately, do sometimes drop out, and if you try and do measures just at the beginning and end you miss out on those clients who have terminated therapy prior to a planned ending. In fact, that can give you better results (because you’re only looking at the outcomes of those who finished properly, who tend to do better) but it’s biased and inaccurate. Session by session monitoring means that you’ve always got a last score for every client, and now most funders or commissioners would expect to see data gathered in that way. If you’ve only got results from 30% of your sample, it really can’t tell you much about the overall picture.

Generally, outcome measures are completed at the start of a session—or before the start of a session—so that clients’ responses are not too affected by the session content. Process measures are generally completed towards the end of a session as they are a reflection on the session itself (but with a bit of time to discuss any issues that might come up).

Analysing the Data

Before you start a service evaluation, you have to know what you are going to do with the data. After all, what you don’t want is to a big pile of CORE-OM forms in one corner of your storage room!

That means making sure you price in to any evaluation the costs, or resources, of inputting the data, analysing it, and writing it up. It simply not fair to ask clients, and therapists, to use hundreds of evaluation forms if nothing is ever going to happen to them.

The good news is that most of the forms, or the sites that the forms come from, tell you how to analyse the data from that form.

The simplest form of analysis, for pre-/post- evaluations, is to look at the average score of clients at the beginning of therapy on the measure, and then their average score at the end. Remember to only use clients who have completed both pre- and post- forms. That will show you whether clients are improving (hopefully) or getting worse.

With a bit more sophisticated statistics you can calculate what the ‘effect size’ is. This is a standardised measure of the magnitude of change (after all, different measures will change by different amounts). The effect size can be understood as the difference between pre- and post- scores divided by the ‘standard deviation’ of the pre- scores (this is the amount of variation in scores, which you can work out via Excel using the function ‘stdev’). Typically in counselling and psychotherapy services, the effect size is around 1, and you can compare your statistics with other services in your field, or with IAPT, to see how your service is doing (although, of course, any such comparisons are ultimately very approximate).

What you can also do is to find out the percentage of your clients that have shown ‘reliable change’ (which is change more than a particular amount, to compensate for the fact that measures will always be imprecise), and ‘clinical change’ (the amount of clients who have gone from clinical to non-clinical bands and vice versa). If you look around on the internet, you can normally find the clinical and reliable change ‘indexes’ for the measures that you are using (though some don’t have them). For the PHQ-9 and GAD-7, you can look here to see both calculations for reliable and clinical change, and the percentages for each of these statistics that were found in IAPT.

Online Services

One way around having to input and analyse masses of data yourselves is to use an online evaluation service. This can simplify the process massively, and is particularly appropriate if you want to combine service evaluation with regular systematic feedback for clinicians and clients. Most of these (though not all) can host a wide range of measures, so they can support the particular evaluation that you choose to develop. However, these services come at a price: a license, even for an individual practitioner, can be in the hundreds or thousands of pounds. Normally, you’d also need to cost in the price of digital tablets for clients to enter the data on.

My personal recommendation for one of these services is:

At the CREST Research Clinic we’ve been using this system for a few years now, and we’ve been consistently impressed with the support and help we’ve received from the site developers. Bill and Tony are themselves psychotherapists with an interest in—and understanding of—how to deliver the best therapy.

Other sites that I would recommend for consideration, but that I haven’t personally used, are:

Challenges

In terms of setting up and running a service evaluation, one of the biggest challenges is getting counsellors and psychotherapists ‘on board’. Therapists are often sceptical about evaluation, and feel that using measures goes against their basic values and ways of doing therapy. Here, it can be helpful for them to hear that clients, in fact, often find evaluation tools quite useful, and are often (though not always) much more positive about it than therapists may assume. It’s perhaps also important for therapists to see the value that these evaluations can have in securing future funding and support for services.

Another challenge, as suggested above, is simply finding the time and person-power to analyse the forms. So, just to repeat, do plan and cost that in at the beginning. And if it doesn’t feel like that is going to be possible, do consider using an online service that can process the data for you.

For the evaluation to be meaningful, it needs to be consistent and it needs to be comprehensive. That means it’s not enough to have a few forms from a few clients across a few sessions, or just forms from assessment but none at endpoint. Rather, whatever you choose to do, all therapists need to do it, all of the time. In that respect, it’s better just to do a few things well, rather than trying to overstretch yourself and ending up with a range of methods done patchily.

Some ‘Template’ Evaluations

Finally, I wanted to suggest some examples of what an evaluation design might look like for particular aims, populations, and budgets:

Aim: Showing evidence of effectiveness to the external world. Population: adults with range of difficulties. Budget: minimal

  • CORE-10: Assessment, and every session

  • CORE Assessment Form

  • Analysis: Service usage statistics; pre- to post- change, effect size, % reliable and clinical change

Aim: Showing evidence of effectiveness to the external world, enhancing outcomes. Population: young people with range of difficulties. Budget: minimal

  • YP-CORE: Assessment, and every session

  • Current View: Assessment

  • ESQ: End of therapy

  • Analysis: Service usage statistics; pre- to post- change, effect size, % reliable and clinical change; satisfaction (quantitative and qualitative analysis)

Aims: Showing evidence of effectiveness to the external world, enhancing outcomes. Population: adults with depression. Budget: medium

  • PHQ-9: Assessment and every session

  • CORE Assessment Form

  • Helpful Aspects of Therapy Questionnaire

  • Patient Experience Questionnaire: End of Therapy

  • Analysis: Service usage statistics; pre- to post- change, effect size, % reliable and clinical change; helpful and unhelpful aspects of therapy (qualitative analysis); satisfaction (quantitative and qualitative analysis)

And finally…

Please note, the information, materials, opinions or other content (collectively Content) contained in this blog have been prepared for general information purposes. Whilst I’ve endeavoured to ensure the Content is current and accurate, the Content in this blog is not intended to constitute professional advice and should not be relied on or treated as a substitute for specific advice relevant to particular circumstances. That means that I am not responsible for, nor will be liable for any losses incurred as a result of anyone relying on the Content contained in this blog, on this website or any external internet sites referenced in or linked in this blog. I also can’t offer advice on individual evaluations. Sorry… but hope the information here is useful.

Publishing your research: Some pointers

Why bother?

Let’s say this, up front: it’s hard work getting your research published. It’s rarely just a case of cutting and pasting a few bits of your thesis, or reformatting an SPSS table or two, and then sending it off to the BMJ for their feature article. So before you do anything, you really need to think, ‘Have I got the energy to do it?’ ‘Do I really want to see this in print?’ And being clear about your reasons may give you the motivation to keep going when every part of you would rather give up. So here’s five reasons why you might want to publish your research.

  1.  If you want to get into academia, it’s pretty much essential. It’s often, now, the first thing that an appointment panel will look at: how many publications you have, and in what journals. 

  2. Even if your focus is primarily on practice, a publication can be great in terms of supporting your career development. It can look very impressive on your CV—particularly if it’s in an area you’re wanting to develop specialist expertise in. Indeed, having that publication out there establishes you as a specialist in that field, and that can be great in terms of being invited to do trainings, or teaching on courses, or consultancy.

  3. It’s a way of making a contribution to your field—and that’s the very definition of doctoral level work. You’ve done your research, you’ve found out something important, so let people know about it. If you’ve written a thesis, it may just about be accessible to people somewhere in your university library, but they’re going to have to look pretty hard. If it’s in a journal, online, you’re speaking to the world.

  4. …And that means you’re part of the professional dialogue. It’s not just you, sitting in your room, talking to your cat: you’re exchanging ideas and evidence with the best in the field—learning, as well as being learnt from.

  5. You owe it to your participants. For me, that’s the most important reason of all. Your participants gave you their time, they shared with you their experiences—sometimes very deeply.  So what are you going to do with that? Are you just going to use it to get your award—for your own private knowledge and development; or are you going to use it to help improve the lives of the people that your participants represent? In this sense, publishing your work can be seen as an ethical responsibility.  

Is IT good enough?

Yes. Almost certainly. If it’s been passed, at Master’s level and especially at doctoral, it means, by definition, that it’s at a good enough standard for publication somewhere. It’s totally understandable to feel insecure or uncertain about your work—we all can have those feelings—but the ‘objective’ reality is that it’s almost certainly got something of originality, significance, and rigour to contribute to the public domain.

Focus

If you’ve written a thesis—and particularly a doctoral one—you may have been covering several different research questions. So being clear about what you want to focus on in your publication, or publications, may be an important next step. Get clear question(s), and be clear about the particular methods and parts of your thesis that answer them. That means that some of your thesis has to go. Yup, that’s right: some of that hard fought, painful, agonised-over-every-word-at-four-in-the-morning will have to be the mercy of your Delete key. That can be one of the hardest parts of converting your thesis to a publication—it’s a grieving process—but it’s essential to having something in digestible form for the outside world.

And, of course, you may want to try and do more than one publication. For instance, you might report half of your themes in one paper, and then the other half in another paper; or, if you did a mixed methods study, you could split it into quant. and qual. Or you might divide your literature review off into a separate paper, or do a focused paper on your methodology. ‘Salami slicing’ your thesis too much can end up leaving each bit just too thin, but if there’s two or more meaningful chunks that can come of your work, why not? 

Finding the right journal

This is one of the most important parts of writing up for publication, and easily overlooked. Novice researchers tend to think that, first, you do all your research, write it up for publication, and then only at the end do you think about who’s going to publish it. But different journals have different requirements, different audiences, and publish different kinds of research; so it’s really important to have some sense of where you might submit it to long before you get to finishing off your paper. That means you should have a look at different journal website, and see what kinds of papers they publish and who they’re targeted towards—and take that into account when you draft your article. 

Importantly, each journal site will have ‘Author Guidelines’ (see, for instance, here) and these are essential to consult before you submit to that journal. To be clear, these aren’t a loose set of recommendations for how they’d like you to prepare your manuscript. They’re generally a very strict and precise set of instructions for the ways that they want you to set it out (for instance, line spacing, length of abstract), and if you don’t follow them, you’re likely to just get your manuscript returned with an irritated note from the publishing team. Particularly important here is the length of article they’ll accept. This really varies across journals, and is sometimes by number of pages (typically 35 pages in the US journals), sometimes by number of words (generally around 5-6,000 words)—and may be inclusive of references and tables, etc., or not. So that’s really important to find out before you submit anywhere, as you may find out that you’re thousands of words over the journal’s particularly limit. Bear in mind that, particularly with the higher impact journals (see below), they’re often looking for reasons to reject papers. They’re inundated: rejecting, maybe, 80% of the papers submitted to them. So if they don’t think you’ve bothered to even look at their author guidelines, they may be pretty swift in rejecting your work.  

So which journals should you consider? There’s hundreds out there and it can feel pretty overwhelming knowing where to start. One of the first choices is whether to go with a general psychotherapy and counselling research journal, or whether something more specific to the field you’re looking at. For instance, if your research was on the experiences of clients with eating disorders in CBT, you could go for a specialised eating disorders journal, or a specialised CBT journal, or a more general counselling/psychotherapy publication. This can be a hard call, and generally you’re best off looking at the journal sites, as above, to see what kind of articles they carry and whether your research would fit in. 

Note, a lot of psychotherapy and mental health journals don’t publish qualitative research, or only the most positivist manifestations of it (i.e., large Ns, rigorous auditing procedures, etc.). It’s unfortunate, but if you look at a journal’s past issues (on their site) and don’t see a single qualitative paper, you may be wasting your time with a qualitative submission: particularly if it’s underlying epistemology is right at the constructionist end of the spectrum. And, if you’re aiming to get your qualitative research published in one of the bigger journals, it’s something you may want to factor in right at the start of your project: for instance, with a larger number of participants, or more rigorous procedures for auditing your analysis.

You should also ask your supervisor, if you have one, or other experienced people in the field, where they think you should consider submitting to. If they’ve worked in that area for some time, they should have some good ideas.  

Impact factor

Another important consideration is the journal’s impact factor. This is a number from zero upwards indicating, essentially, how prestigious the journal is. There’s an ‘official’ one from the organisation Clarivate; but these days most journals will provide their own, self-calculated impact factor if they do not have an official one. You can normally find the impact factor displayed on the journal’s website (the key one is the ‘two year’ impact factor—sometimes just called the ‘impact factor’—as against the five year impact factor). To be technical, the impact factor is the amount of times that the average article in that journal is cited by other articles over a particular period: normally two years. So the bigger the journal’s impact factor, the more that articles in that journal are getting referenced in the wider academic field—i.e., impact. The biggest international journals in the psychotherapy and counselling field will have an impact factor of 4 or 5, and ones of 2 or 3 are still strong international publications. Journals with an impact factor around 1 may tend towards a national rather than international reach, and/or be at lower levels of prestige, but still carrying many valuable articles. And some good journals may not have an official impact factor at all: journals have to apply for an official one and in some cases the allocation process can seem somewhat arbitrary.

Of course, the higher the journal’s impact factor, the harder it is to get published there, because there’s more people wanting to get in. So if you’re new to the research field, it’s a great thing to get published in a journal with any impact factor at all; and you shouldn’t worry about avoiding a journal just because it doesn’t have an impact factor, or if it’s fairly low. At the same time, if you can get into a journal with an impact factor of 1 or above that’s a great achievement, and something that’s likely to make your supervisor(s), if they’re co-authors on the paper (see below), very happy. For more specific pointers on publishing in higher impact journals, see here.

These days, the impact of a journal may also be reported in terms of its quartile: so from Q1 to Q4.  Essentially Q1 journals are those with impact factors within the top 25% of the subject area, and down to Q4 journals which are in the lowest 25%.  

In thinking about impact factor, a key question to ask yourself is also this: Do I want to (a) just get something out there with the minimum of additional effort, or (b) try and get something into the best possible journal, even if it takes a fair bit of extra work. There’s no right answers here: if you have got the time, it’s great if you can commit to (b), but if that’s not realistic and/or you’re just sick and tired of your thesis, then going for (a) is far better than not getting anything out at all.

General counselling and psychotherapy research journals

If you’re thinking of publishing in a general therapy research journal, one of the most accessible to get published in is Counselling Psychology Review – particularly if your work is specific to counselling psychology.  The word limit is pretty restrictive though. There’s also the European Journal for Qualitative Research in Psychotherapy, which is specifically tailored for the publication of doctoral or Master’s research, and aims to ‘provide an accessible forum for research that advances the theory and practice of psychotherapy and supports practitioner-orientated research’. If you’re coming from a more constructionist perspective, a journal like the European Journal of Psychotherapy & Counselling might also be a good first step, which publishes a wide range of papers and perspectives.

For UK based researchers, two journals that are also pretty accessible are Counselling and Psychotherapy Research (CPR) and the British Journal of Guidance and Counselling (BJGC). Both are very open to qualitative, as well as quantitative studies; and value constructionist starting points as well as more positivist ones. The editors there are also supportive of new writers, and know the British counselling and psychotherapy field very well. See here for an example of a recent doctoral research project published in the BJGC (Helpful aspects of counselling for young people who have experienced bullying: a thematic analysis), and here for one in CPR (Helpful and unhelpful elements of synchronous text‐based therapy: A thematic analysis).

 Another good choice, though a step up in terms of getting accepted, is Counselling Psychology Quarterly. It doesn’t have an official impact factor, but it has a very rigorous review process and publishes some excellent articles: again, both qualitative and quantitative.

Then there’s the more challenging international journals, like Journal of Clinical Psychology, Psychotherapy Research, Psychotherapy, and Journal of Counseling Psychology, with impact factors around 3 to 5 (in approximate ascending order). They’re all US-based psychotherapy journals, fairly quantitative and positivist in mindset (though they do publish qualitative research at times), and if you can get your research published in there you’re doing fantastically. Like a lot of the journals in the field, they’re religiously APA in their formatting requirements, so make sure you stick tightly to the guidelines set out in the APA 7th Publication Manual. A UK-based equivalent of these journals, and open-ish to qualitative research (albeit within a fairly positivist frame), is Psychology and Psychotherapy, published by the BPS.

There’s even more difficult ones, like the Journal of Consulting and Clinical Psychology with an impact factor of 4.5, and The Lancet is currently at 53.254.  But the bottom line, particularly if you’re a new researcher, is to be realistic. Having said that, there’s no harm starting with some of the tougher journals, and seeing what they say. At worse, they’re going to reject your paper; and if you can get to the reviewing stage (see below), then you’ll have a really helpful set of comments on how to improve your work. 

If a journal requires you to pay to publish your article, it’s possible a predatory publisher (‘counterfeit scholarly publishers that aim to trick honest researchers into thinking they are legitimate’, see APA advice here). In particular, watch out for emails, once you’ve completed your thesis, telling you how wonderful your work is and how much they want to publish it in their journal—only to find out later that they charge a fortune for it. You may also find yourself getting predatory requests to present your research at conferences, with the same underlying intent. Having said that, an increasing range of reputable journals—particularly online ones that publish papers very quickly, like Trials—do ask authors to pay Article Processing Charges (APC). Generally, you can tell the ‘kosher’ ones by their impact factor and whether they have a well-established international publisher. It’s also very rare for non-predatory journals to reach out to solicit publications. Check with a research supervisor if you’re not sure, but be very, very wary of handing over any money for publication.  

Writing your paper

So you know what you’re writing and who for, now you just have to write it. But how do you take, for instance, your beautiful 30,000 word thesis and squash it down to a paltry 6,000 words?

If you’re trying to go from thesis to article, the first thing is that, as above, you can’t just cut and paste it together. You need to craft it: compiling an integrated research report that is carefully knitted together into a coherent whole. It’s an obvious thing to say, but the journal editors and reviewers won’t have seen your thesis, and they’ll care even less what’s in it. So what they’ll want is a self-contained research report that stands up in its own right—not referring back to, or in the context of, something they’ll never have time to read. That’s particularly important to bear in mind if you’re writing two or more papers from your research: each needs to be written up as a self-contained study, with its own aims, methods, findings, and discussion.

In writing your paper, try and precis the most important parts of your thesis in relation to the question(s) that you’re asking. Take the essence of what you want to say and try and convey it as succinctly and powerfully as possible. Think ‘contracting’ or ‘distilling’: reducing a grape down to a raisin, or a barley mash down to a whiskey—where you’re making it more condensed but retaining all the goodness, sweetness, and flavour. That doesn’t mean you can’t cut and paste some parts of your thesis into the paper, but really ask yourself whether they can be condensed down (for instance, do you really need such long quotes in your Results section?), and make sure you write and rewrite the paper until it seamlessly joins together.

Your Results are generally the most important and interesting part of your paper, so often the part you’ll want to keep as close to its original form as possible. So if you’ve got, say, 7,000 words for your paper, you may want your Results to be 2-3,000 of that (particularly if it’s qualitative). Then you can condense everything else down around it. Your Introduction/Literature Review may be reducible to, perhaps, 500-1,000 words. Maybe 1,000 words for your Methods and Discussion sections; 1,000 words for References. 

If you’ve written a thesis, you may be able to cut some sections entirely. If you’re submitting to a more positivist journal, your reflexivity section can often just go; equally your epistemology. Sorry.  If your study is qualitative, you may also find that you can cut down a lot of the longer quotes in your Results. Again, try and draw out the essence of what you are trying to say there… and just say it.

Generally, and particularly for the higher-end US journals, you’re best off following the structure of a typical research paper (and often they require this): Background, Method, Results, Discussion, References. They’re may be more latitude with the more constructionist journals but, again, check previous papers to see how research has been written up. 

Make sure you write a very strong Abstract (and in the required format for the journal). It’s the first thing that the editor, and reviewers, will look at; and if it doesn’t grab their attention and interest then they may disengage with the rest. There’s some great advice on writing abstracts in the APA 7th Publication Manual as well as on the internet (for instance, here).  

Supervisors and consultants

If you’ve had a supervisor, or supervisors, for your research work, there’s a question of how much you involve them in your publication, and whether you include them as co-author(s). At many institutions, there’s an expectation that, as the supervisor(s) have given intellectual input into the research, they should be included as co-author(s), though normally only as second or third in the list. An exception to the latter might be if a student feels like they don’t want to do any more work at all after they’ve submitted their thesis, in which case there might be an agreement that one of the supervisors take over as first author. Here, as with any other arrangement, the important bit is that it’s agreed up front and everyone is clear about what’s involved. 

Just to add, as a student, you should never be pressurised by a supervisor into letting them take the first author role. I’ve never seen this actually happen, but have heard stories of it; and if you feel under any coercion at all then do talk to your Course Director or another academic you trust.

The advantage of keeping your supervisor(s) involved is that they can then help you with writing up for publication, and that can be a major boost if they know the field and the targeted journal well. So use them: probably, the best way of getting an article published in a journal is by co-authoring it with someone who’s already published there. A way that it might work, for instance, is that you have a first go at cutting down your thesis into about the right size, and then the supervisor(s) work through the article, tidying it up and highlighting particular areas for development and cutting. Then it comes back to you for more work, then back to your supervisor(s) for checking, then back to you for a final edit before you submit.  

One final thing to add here: even though you may be working with people more senior and experienced to you, if you are first author on the paper, you need to make sure you ‘drive’ the process of writing and revising, so that it moves forward in a timely manner. So, for instance, if one of your supervisors is taking a while to get back to you, email them to follow up and see what’s happening; and make sure you always have a sense of the process as a whole. This can be tough to do, given the power relationship that would have existed if you were their supervisee; but, in my experience, the most common reason that efforts at publication fizzle out are because there’s no one really ‘holding’ or driving the process: no one making sure it does happen. Things fall through gaps: a supervisor doesn’t respond for a month or two, no one follows them up, the other supervisor wanders off, the student gets on with other things… So spend a bit of time, at the start, agreeing who’s going to be in charge of the process as a whole (normally the first author) and what roles other authors are going to have. And, if it’s agreed that you are in the driving seat, you’ve got both the right and the responsibility to follow up on people to make sure it all gets done.

How do you submit?

That takes us to the process of submitting to a journal.  So how does it work? Nearly all journals now have an online submission portal so, again, go to the journal website and that will normally take you through what you need to do. Submission generally involves registering on the site, then cutting and pasting your title and abstract into a submission box, entering the details of the author(s) and other key information, and uploading your papers. The APA 7th Manual has some great advice on how to prepare your manuscript so that it’s all ready for uploading (or see here), and if you follow that closely you should be ok for most journals.

You also normally need to upload a covering letter when you submit, which gives brief details of the paper to the Editor. This can also cover more ‘technical’ issues, like whether you have any conflicts of interest (have you evaluated, for instance, an organisation that you’re employed by?), and confirmation of ethical approval. If you’ve submitted, or published, related papers that’s also something you can disclose here. Generally, it’s fine to submit multiple papers on different aspects of your thesis, but they should be different; and it’s always good just to let the editor know so that it doesn’t come as a surprise to them later. 

Note, you definitely mustn’t submit the same paper (or similar papers) to more than one journal at any one time. That’s a real no-no. Of course, if your paper gets rejected it’s fine to try somewhere else (see below), but you could get into a horrible mess if you submitted to more than one journal in parallel (for instance, what happens if they both accept it?). So most journals ask you, on submission, to confirm that that’s the only place you’ve sent it to and that’s really important to abide by.  

What happens then?

The first thing that normally happens is that a publishing assistant will then have a quick look over your article to check that it’s in the right format. As above, they can be pretty pernickety here, and if you’re over the word limit, or not doing the right paragraph spacing, or even indenting your paragraphs when you shouldn’t, you can find your article coming back to you asking for formatting changes before it can be considered. So try and get it right first time.

Then, when it’s through that, it’s normally reviewed by the journal editor, or a deputised ‘action editor’. Here, they’re just getting a sense of whether the article is right for the journal, and at about the required level. Often papers will get rejected at that point (a desk rejection), with a standard email saying that they get a lot of submissions, they can’t review everything, it’s no comment on the quality of the paper, etc., etc. Pretty disappointing—and generally not much more feedback than that. Ugh!

If you don’t hear from the journal a week or so after submission, it generally means it’s then got through to the next stage, which is the review process. Here, the editor will invite between about two and four experts in the field to read the paper, and give their comments on it. This process is usually ‘blind’ so they won’t know who you are and you won’t know who they are. In theory, this helps to keep the process more ‘objective’: the reviewers aren’t biased by knowing who you actually are, and they don’t have to worry about ‘come back’ if they give you a bad review.  

The review process can take anything between about three weeks and three months. You can normally check progress on the journal submission website, where it will say something like ‘Under review.’ If it gets beyond three months or so, it’s not unreasonable to write to the journal and ask them (politely) how things are going. But there’s no relationship between the length of the time of the review and the eventual outcome—it’s normally just that one of the reviewers is taking too long getting back to them, and they may have had to look elsewhere. Note, even if it is taking a long time and you’re getting frustrated, you can’t send the paper off somewhere else until things are concluded with that first journal. You could withdraw the paper, but that’s fairly unusual and mostly people wait until the reviews are eventually back.

The ‘decision letter’

Assuming the paper has gone off for review, you’ll get a decision letter email from the editor. This is the most exciting—but also the most potentially heartbreaking part—of the publication process: a bit like opening the envelope with your A-level results in. Generally, this email gives you the overall decision about acceptance/rejection, a summary from the editor of comments on your paper, and then the specific text of the reviewers’ comments.

In terms of the decision itself, the best case scenario is that they just accept it as it is. But this is so rare, particularly in the better journals, that if you ever got one (and I never have), you’d probably worry that something had gone wrong with the submission and review process.

Next best is that they tell you they’re going to accept the paper, but want some revisions. Here, the editor will usually flag up the key points that they want you to address, and then you’ll have the more specific comments from the reviewers. Sometimes, journals will refer to these as ‘minor revisions’, as opposed to more ‘major revisions’, but often they don’t use this nomenclature and just say what they’d like to see changed. Frequently, they don’t even say whether the paper has been accepted or not—just that they’d like to see changes before it can be accepted—and that can be frustrating in terms of knowing exactly where you stand. Generally, though, if they don’t explicitly use the ‘r’ word (‘reject’), it’s looking good.

Then you can get a ‘reject and resubmit’. Here, the editor will say something like, ‘While we can’t accept/have to reject this version of the paper, due to some fairly serious issues or reservations, we’d like to invite you to resubmit a revision addressing the points that the reviewers have raised’. In my experience, about 60% of the time when you resubmit a rejected paper you eventually get it through, and about 40% of the time they subsequently reject it anyway. The latter is pretty frustrating when you’ve done all that extra work, but at least you’ve had a chance to rework the paper for a submission elsewhere. 

Then, there’s a straight rejection, where the editor says something fairly definitive like, ‘…. your paper will not be published in our journal.’ That’s pretty demoralising but, at least, if you’ve got to this stage, you’ve nearly always got some very helpful feedback from experts in this field to help you improve your work.

Emotionally, the editorial and reviewing feedback can be pretty bruising, especially when it’s a rejection. Reviewers don’t tend to pull punches: they say what they think—particularly, perhaps, because they’re under the cover of anonymity. So you do need to grow a fairly thick skin to stick with it.  Having said that, a good reviewer should never be diminishing, personal, or nasty.  Even when rejecting a submission, they’ll be able to highlight strengths as well as limitations, and to encourage the author to consider particular issues and pursue particular lines of enquiry, to make the best of their work and their own academic growth. So if something a reviewer says is really hurtful, it’s probably less about the quality of your work, and more about the fact that they’re being an a*$e (at least, that’s what I tell myself!).

Most journals do have some kind of appeal process if you’re really unhappy with the decision made. But you need a good, procedural argument for why you think the editorial decision was wrong (for instance, that it was totally out of step with the actual reviews, or that the reviewers hadn’t actually read your paper) and, in my experience, appeals don’t tend to get too far. However, I have heard of one or two instances of successful outcomes.

By the way, sometimes, quite quickly after you’ve started to submit papers (and possibly even before), you may be asked to review for the journal yourself. That can be a great way of getting to know the reviewing process better—from the other side. It’s also part of giving back to the academic community: if people are spending time looking at your work, it’s only fair you do the same. So do take up that opportunity if you can. There’s some very helpful reviewer guidelines here.  

Revising and resubmitting

If you’re asked to make revisions, journals will generally give you six months or so—less if they’re relatively minor. Here, it’s important to address every point raised by each of the reviewers. That doesn’t mean you have to do everything they ask for, but you do have to consider each point seriously, and if you disagree with what they’re saying, you need to have a good reason for it. Generally, you want to show an openness to feedback and criticism, rather than a defensive or a closed-minded attitude. If the editor feels like they’re going to have to fight with you on each point, they might just reject the paper on resubmission.

As well as sending back the revised papers, you’ll need to compile a covering letter indicating how you addressed each of the points that the reviewers’ raised. You may want to do this as a table as you go along: copy-pasting each of the reviewers’ points, and then giving a clear account of how you did—or why you did not—respond to that issue.

Pay particular attention to any points flagged up by the editor. Ultimately it will be their decision whether or not to accept your paper, so if they’re asking you to attend to some particular issues, make sure you do so. 

Resubmissions go back through the online portal. If the changes required are relatively minor, it may just be the editor looking over them; anything more substantive and they’ll go back to the reviewers again for comment. Bear in mind that the reviewers are often the original ones who looked at your paper, so ignore their comments at your peril.

It’s not unusual to have three or four rounds of this review process: moving, for instance, from a ‘revise and resubmit’ to ‘major revisions’ to ‘minor changes’. At worst, it can feel petty and irritating; but, at best, and far more often, it can feel like a genuine attempt by your reviewers to help you improve the paper as much as possible. The main thing here is just to be patient and accept that the process can be a lengthy one. If you’re in a rush and just desperate to get something out whatever it’s quality, you’re likely to be profoundly frustrated—unless you’re prepared to accept publication in a journal of much lower quality.  

Once it’s accepted

Yay! You got there! That’s it… not quite. It’s brilliant to have that final acceptance letter from the journal telling you that they’ll now go ahead and publish your paper, but there is still a little more to do. A few weeks after the acceptance email, they’ll send you a link to a proof of the paper, where there’ll be various, relatively minor copy-editing corrections and queries. For instance, they may suggest alternate wording for sentences they think could be improved, or ask you to provide the full details for a reference. Sometimes, this may be in two stages: with, first, a copy-edited draft of your manuscript, and then a fully formatted proof). Note, at this point, they really don’t like you to make any substantive changes, so anything you want to see in the final published article should be there in your final submitted draft.

Then that it is. Normally the paper will be out, online, a week or so after that. And once it is, you can finally celebrate, but do also make sure you let people know about the paper, and give everyone the link via social media. The journal, itself, are unlikely to do any specific promotion of the article, so it’s up to you to tell colleagues about it and encourage them to let others in the field know.

Open Access?

Although it’s great you’ve got your paper out, the final pdf version may only be available to people who have access to the journal. So students at higher education institutes are likely to be fine, as are colleagues working for large organisations like the NHS, but what about counsellors or psychotherapists who don’t have online access, and where the cost of purchasing single articles are often prohibitively high? One possibility is that you (or the institution you are affiliated to) can pay to make your article ‘open access’. However, this can cost £1000s (unless the University has a pre-established agreement with the publisher) and is not something most of us can afford.

Fortunately, journals normally allow you to post either your original submission to the journal (an ‘author’s original manuscript’, or ‘preprint’ version of your article), or your final submission (a ‘prepublication’, ‘author final’, ‘postprint’, or ‘author accepted manuscript’ version of your article) on an online research depository, such as ResearchGate. Policies vary, so check the specific policies for the journal that published your paper:

This version of your paper won’t be the exact article that you published, and it won’t have the correct pagination etc., but if you prepare it well (see an example, here), then it means that those who don’t have access to journal sites can still find, read, and cite your research. Different journals do have different policies on this, though, so make sure you check with the specific publisher of your journal before making any version of your paper publicly available. Generally, what the publishers are very vigilant about is the making available, in a public place, of the final formatted pdf of your paper (unless, as above, it’s specifically open access).

Trying elsewhere

If your paper gets rejected, your choices are (a) just to give up, (b) resend the paper as is it somewhere else, or (c) make revisions based on the feedback and then resubmit elsewhere. There’s also, of course, a lot of grey areas between (a) and (b) depending on how many changes you feel willing—and able—to make. Generally, if you can learn from the feedback and revise your paper that’s not a bad thing, and can help form a stronger submission for next time. Of course, it is always possible that the next set of reviewers will see things in a very different way; and sometimes changes made to address one set of concerns will then be picked on by the next set of reviewers as problems in themselves. As for (a), well, I promise you this: if the research is half-decent, then you can always get it published somewhere. Bear in mind that, as above, if you’ve been awarded a doctorate for your research (and, to some extent, a Master’s), it’s publishable by definition

Generally, when people get their papers rejected, they move slowly down the impact hierarchy: so to journals that might be more tolerant of the ‘imperfections’ in your paper. But there’s no harm in trying journals at a similar level of impact when you’re trying somewhere else or even higher up—particularly when you really don’t agree with the rejecting journal’s feedback.

Ultimately, it’s about persistence. To repeat: if you want to get something published, and it’s passed at doctoral (or, often, Master’s) level, you will. But it needs resilience, responsiveness, and a willing to put up with a lot of knockbacks.   

Other pathways to impacts

Journals aren’t the only place where you can get your research out to a wider audience and make an impact. For instance, you could write a synopsis of your thesis and post it online: such as on Researchgate. You won’t get as big a readership as in an established journal, but at least it will be more accessible than your university library, and you can tell people about it via social media. Or you could do a short blog about your research, or make a video, or talk to practitioners and other stakeholders about your work. If you want to make your research findings widely accessible to practitioners, you could also write about them for one of the counselling and psychotherapy magazines, like BACP’s Therapy Today or BPS’s The Psychologist.   

There’s also many different conferences that you can go to to present your findings: as an oral paper, or simply as a poster. Two of the best, for general counselling and psychotherapy research in the UK, are the annual research conference of the British Association for Counselling and Psychotherapy (BACP), and the annual conference of the BPS Division of Counselling Psychology (DCoP). Both are very friendly, encouraging, and supportive; and you’ll almost certainly receive a very warm welcome just for having the courage to present your work. At a more international level is the annual conference of the Society for Psychotherapy Research (SPR). That’s a great place to meet many of the leading lights in the psychotherapy research world, and is still a very friendly and supportive event. 

You can also think about ways in which you might want your work to have a wider social and political impact. Would it make sense, for instance, to send a summary to government bodies, or commissioners, or something to talk to your local MP about?

Of course, this could all be in addition to having a publication (rather than instead of it), but the main point here is that, if you want your research to have impact, it doesn’t just have to be through journal papers.  

To conclude…

When you’ve finished a piece of research—and particularly a long thesis—often the last thing you’ll want to be doing is reworking it into one or more publications. You can’t stand the sight of it, never want to think about it again—let alone take the research through a slow and laborious publication process. But the reality is, as people often say, the longer you leave it the harder it gets: you move away from the subject area, lose interest; and if you do want to publish at a later date, you’ll have to familiarise yourself with all the latest research (and possibly without a library resource to do so). So why not just get on with it, get it out there; and then you can have your work, properly, in the public domain, and people can use it and learn from it, and improve what they do and how they do it. And then, instead of spending the next few decades wishing you had done something with all that research, you can really, truly, have the luxury of never having to think about it again.

Acknowledgements

Many thanks to Jasmine Childs-Fegredo, Mark Donati, Edith Steffen, and trainees on the University of Roehampton Practitioner Doctorate in Counselling Psychology for comments and suggestions. 

Further Resources

There is a great, short video here from former University of Roehampton student, Dr Jane Halsall, talking about her own experience of going from thesis to published journal paper. Jane concludes, ‘You’re doing something for the field, and you’re doing something for the people who have actually taken the time out to participate. So be encouraged, and do do it.’

An accessible set of tips on publishing in scholarly is also available from the APA:

Disclaimer

The information, materials, opinions or other content (collectively Content) contained in this blog have been prepared for general information purposes. Whilst I’ve endeavoured to ensure the Content is current and accurate, the Content in this blog is not intended to constitute professional advice and should not be relied on or treated as a substitute for specific advice relevant to particular circumstances. That means that I am not responsible for, nor will be liable for any losses incurred as a result of anyone relying on the Content contained in this blog, on this website, or any external internet sites referenced in or linked in this blog.

The Viva: Some Pointers

The following blog is for Master’s or doctoral level students writing research dissertations in the psychological therapies fields. The pointers are only recommendations—different trainers, supervisors, and examiners may see things very differently.

Many thanks to Jasmine Childs-Fegredo, Mark Donati, and Edith Steffen for comments and suggestions.

 ***

For doctoral students, the viva is the endpoint of your academic journey, and can be the most dreaded part.  So what is it, and what should you do to make it go as well as possible?

The Set Up

Typically, you’ll have two examiners: an ‘internal’ (someone based at your university), and an ‘external’ (someone based at another university).  The external usually carries more weight, and may have more influence on the final decision.  You may also have a ‘Chair’ (normally someone based at your university as well), but their role will just be to manage the viva examination.  They don’t have a role in assessing you.

 Often, you can choose whether or not to have your supervisors present at the viva, though they won’t be able to say anything.  This can feel like moral support, and they can take notes on what might need to be revised.  However, you may also feel more pressure having additional people in the room.

Typically, a viva lasts for about 90-120 minutes, though that can vary a lot.  If longer, you’ll normally get a short break.

What viva examiners often do is to go through your thesis chapter by chapter, asking questions and discussing with you aspects of your work along the way.  Sometimes your examiners will take it turns to ask questions, or the external may take more of a lead.  However, this can depend on the examiners’ areas of expertise.  For example, if your external knows more about your methods, and your internal more about your content area, they may divide the questions up in that way.

Prior to the viva, both of your examiners will have read through your thesis, and written an independent report of what they make of it, and approximately what outcome they think you should be awarded.  In the vast majority of cases, this will be either ‘minor amendments’ (for instance, adding more on reflexivity, discussing the limitations in more depth) or ‘major amendments’ (for instance, restructuring the literature review, revising the analysis).  In a small number of cases, they may also feel that you need to collect more data—a very major amendment.  It’s also possible that they’ll feel the thesis should fail but, thankfully, that’s very rare, and something that your supervisors would normally alert you to before you submit.  Equally rare is that the examiners will just pass your thesis without wanting to see any changes at all, so it may be best to go into the viva assuming that the examiners will ask you to make revisions to some degree—even if it’s just correcting typos.  Before they meet you, your examiners will also have met with each other and shared their views on your thesis, coming up with a list of questions to structure the viva by.

Your examiners may start by telling you their overall assessment of your thesis, or they may not.  If this doesn’t happen, don’t read anything into it—some examiners just prefer not to do so.

After they’ve talked your thesis through with you, the examiners will ask you to leave the room (for 30 minutes or so), and then they’ll discuss with each other what they think the outcome should be and what changes they think you should make.  Then they’ll invite you back in and share the result with you.  If, as normal, they’re asking for some amendments, they’ll go through them with you but you won’t need to write them down, as you’ll be sent the feedback in writing soon after the viva.

What’s it For?

As an external examiner, what I’m wanting from the viva is three things.  First, I want to make sure that the student has really written their thesis, and not got someone else in to do it for them.  So that means I’m looking to see that they can talk about their work in a fairly fluent and knowledgeable way.  Second, although I’ll come into the viva with an outcome in mind and some idea of the kinds of amendments that I might want to see, I’m also open to revising that, depending on how the candidate talks about their work.  For instance, I might feel that they should have conduct a systematic literature review rather than a narrative one, but if they present a convincing argument for why they did the latter, then I may be happy to let that go.  Third, I might want to convey—and explain—to the student why I think they should make certain changes to their thesis, and what those changes are.  

 Remember that your examiners, like your supervisors, will almost certainly want to see you get through.  No one wants to fail anyone—we all know how much work a thesis takes.  But we also will want to make sure things are fair: if it feels like you haven’t got your head around certain things, or done the work that you’ve needed to do, then it wouldn’t feel right to pass you alongside others.  And we’re also aware that your thesis will be lodged publicly, for all to access and read.  So we want to see it in the best shape possible: something you can be proud of and that reflects the best of your abilities.

How to Prepare

Before the viva, have a really good few read throughs of your thesis so that you know it well.  You may have completed it several months before the viva, so it’s important to re-familiarise yourself with it—particularly the more tricky or complex parts.

Practice vivas are essential.  Your supervisor(s) will often be willing to do this with you.  If not, or as well as, do practice vivas with your peers or friends.  Get them to ask you questions about your thesis—particularly the more difficult bits (like epistemology, or your choice of methods, or any statistical tests) so you can get practised at talking these elements through.  Talk to your friends, your family, your cat about your thesis (as much as they can bear it) so that you’re really familiar with what you did and why.

What to Take to the Viva?

One of my personal bugbears, as an examiner, is when students come to a viva without a copy of their own thesis, and then have to borrow mine to answer questions.  So make sure you bring yours along, with sections clearly marked so you can find your way around it when asked about different parts.  It’s fine also to bring a notepad so you can write down questions.

Nerves

It can be really scary doing a viva, and your examiners should be well aware of that and sensitive to it—bear in mind that they will have gone through one of their own.  So if you get really nervous at the start or during the viva, it’s normally fine to ask for just a bit of time to compose yourself—there’s no rush. You may even want to let the Chair or examiners know at the start, if you think that will help. 

What Will They Ask Me?

Mostly, the questions that your examiners will ask will be specific to your particular thesis.  As indicated above, typically, they’ll go through it chapter by chapter, and ask you to explain, or elaborate on, specific aspects of your work.  The questions will often be on the areas that they feel might need further work.  However, if they feel that really not much needs to be changed, they may just be asking about particular areas of interest to discuss them with you.  After they’ve asked you about a particular area of issue, they may follow this up with further questions or prompts.  Questions may be fairly general (for instance, ‘Can you explain your choice of analytical method?’) or very specific (for instance, ‘On page 125, you indicate that the p-value was .004, but on page 123 you write that the regression analysis wasn’t significant, can you explain that please’).  There’s also some standard questions that examiners may ask, for instance:

  •  Why did you choose to do this study?

  • How did you go about choosing what literature to look at?

  • What was the underlying epistemology for your research?

  • What was the rationale for your sample?

  • Why did you choose this particular method?  Why not xxx method?

  • What are the implications for counselling/psychotherapy/counselling psychology practice of your thesis?

  • What does your research add to the field?

  • What are the limitations of your study?

  • What was the impact of your personal perspective on the study? Biases?

  • What did you personally learn from the study?

Elaborate, Elaborate, Elaborate

In terms of the actual viva, the main bit of advice I would give candidates is to make sure you really elaborate on your answers.  Of course, you want to stay on track with the particular question you’ve been asked, but don’t be too short or pithy in how you respond.  For a typical viva, the examiners may have prepared, say, 10 questions or so, so you need to talk on each area for, perhaps, 10 minutes; and you don’t want a situation where your examiners are constantly having to pump you for answers.  This is your chance to show your depth and breadth of thinking so, for instance, reflect with the examiners on why you made the choices you did, show how you weighed up different possibilities, talk about the details of what you considered and what you found.  Ultimately, what your examiners want to see is that you can think deeply and richly and complexly about things—rather than that you have reached any single definitive conclusions.  So it’s less about getting it ‘right’, and more about showing all the thinking that has been going on. 

Don’t be Defensive

The other main thing I would say is not to be too defensive when you respond to the examiners’ questions and prompts.  As indicated above, they’ll have a view on what they may want you to revise in your thesis, and while you may be able to change their minds to some extent, you don’t want to come across as too rigid or stubborn in your thinking.  If, when they point something out to you, you think, ‘Actually, they’re probably right,’ that should be fine to say, and better than trying to defend something that you can clearly see is in need of adjustment.  Of course, if you think you’re right, do say it and say why, but you don’t have to defend to the bitter end every element of your work.  Better to show, like all of us, that you can sometimes get things wrong and that you’re open to learning and improving.

Be the Expert You Are

As Mark Donati, Director on our Doctorate in Counselling Psychology at the University of Roehampton suggests, don’t be afraid to express your opinion and say what you really think.  Of course, it’s best if this is based on the available evidence; but sometimes the evidence just isn’t available, and then the examiners may be really interested in your ‘best guess’ of what’s going on.  Remember that you are the expert in the area now.  That’s right, you are.  And the examiners may be really excited to hear from you what the view is from the leading edge of the field. 

Don’t Shame your Examiners

That might sound strange to say, but bear in mind that your examiners are also in a social situation, and may be experiencing their own pressures to ‘perform’.  Dr X, for instance, has come down from University Y, and it’s the first time they’ve met your internal examiner Professor Z, whose work they’ve always admired, as well as Chair W, who they don’t know very well but who seems an important figure.  So Dr X wants to show that they’ve got a good understanding of your work, with some intelligent questions to ask, and some good insights about the field.  What that means is, if you want to keep your examiners ‘on side’, treat them with respect and show an interest in what they’re saying and the questions they asked.  You really don’t want to respond to Dr X in a way that may make them feel foolish in front of Professor Z, or like they have to defend themselves.  What this also means is that some of what goes on in the room may not be about you, but also about the dynamics between the rest of them. 

Enjoy

It’s easier said than done, but if you can enjoy your viva (and many students do end up doing so) then that’s great.  Think of it this way: you’ve got a captive audience for two hours who you can talk to about all the work you’ve been doing for the last few years.  And now you’re the expert, so make the most of it: tell them about what you’ve been really thinking, and about some of the complex challenges doing the thesis, and about all your ideas about where the research should go for the future.  It’s your chance to shine, and if you can really connect with your energy and enthusiasm for your work, your examiners are sure to appreciate that—and so might you.

Publishing a Therapy Book: Some Pointers

So you’ve got a great idea for a book in the counselling and psychotherapy field. You’re all excited. You want to write. What do you do next to turn your idea into a fully-fledged publication? 

Who’s it for?

That’s great you’ve got an idea. But a lot of what you need to do is to turn things on their head and ask yourself, ‘Who’s going to want this book?’ Unfortunately, all the excitement and passion we can feel inside ourselves doesn’t necessarily translate into a viable book for a publisher. Their first question is going to be, ‘Who’s going to want to buy this?’ So you really need to be clear about that. Is it for trainees? Is it for practitioners? On person-centred courses? On integrative courses? And you need to be realistic here. Bear in mind that people have hundreds of books on their ‘to buy’ list, so why would they want to buy yours? A book that is targeted towards trainees is likely to be particularly appealing to publishers, because that tends to be their biggest market. And if it’s the kind of thing that would be a core text on a module reading list, bingo, that’s exactly the kind of thing that many publishers will be looking for. 

What else is out there?

You need to know the field. What other kind of books are like it? If there’s something out there similar, that doesn’t necessarily mean that yours is a no goer, but you need to make it clear to the publishers what the unique selling point (USP) of your book is going to be. Maybe it’s more accessible than the previous texts. Maybe it’s for work with children rather than adults? But you need to clearly state to the publishers why your book will fill a gap in the market that isn’t currently filled. And that means more than just quoting what you’re already aware of. It means doing some research on sites like Amazon or Google to have a really good rummage around to find out what’s out there so far.  

What have you written before?

Publishers will want to be reassured that you can write. If you’ve written articles or journal papers before, that’s great; and a book or two will really convince a publisher that you’re going to produce what you say you will. If you’ve never written before, a book is a tough place to start, and you’re probably better off writing and submitting a few articles first—say for Therapy Today—to get a sense of how you feel about writing and what kind of feedback you get. Anyone, I’m sure, can write brilliantly, but believing we can write brilliantly isn’t the same as actually doing it. It requires the ability to be able put things in clear and succinct ways. And, more than anything else I think, it requires the kind of dogged, slightly OCD personality that is determined to go on and on even when you’re exhausted and tired and just want a glass of wine and sleep. If you’re not sure that’s you, then best to find out first. 

co-authored and edited BOOKS

Writing a book with one or more other colleagues can be a great way to take a project forward: not only do you split the work, but you can get to have some great dialogues along the way. The obvious thing to say, though, is to make sure that you really do get along and you’ve agreed the basics of who’s doing what, etc. You really don’t want to get halfway through the book and discover that you’ve got completely different ideas about how it should end up, or your co-author’s moved to Goa and wants to spend the rest of their life doing yoga instead of writing.

You may also be thinking about putting together an edited collection of chapters on a particular topic. Again, that does split the work and means that you don’t have to know everything yourself; but don’t underestimate the effort of identifying, then editing, a whole series of chapters—and liaising with 10-20 authors along the way. Sometimes, when I’ve done that, it’s felt like it would have been easier to write the whole thing myself! Also, publishers don’t tend to like edited books as much as single or co-authored texts. They’re not usually as coherent, or as flowing, and generally they don’t sell as well. So if you do go down that route, I’d suggest taking a strong editorial lead, to make sure that everyone is writing to the same brief and same overall aims.

Which publisher?

There’s lots of different publishers out there in the counselling and psychotherapy field and you’ll need to decide which one to approach first. It’s ‘bad form’ to approach more than one publisher at any point in time, so you’ll need to start with one and, if they don’t like it, go on to another, etc. To find the right publisher, have a look at similar books in the field and see who they are published by. You may want to start there. If it’s a general counselling or psychotherapy textbook, particularly for trainees, Sage might be a great place to start. If it’s a bit more specialised, and particularly related to person-centred therapy or critical perspectives, PCCS Books could a very good choice. Routledge have a very wide ranging list and tend to publish a bit more academic, and specialised, books than Sage. And then there’s many others—like Palgrave, Open University Press, Oxford University Press—all with their own areas of focus and speciality. If you’re not sure, just go to their websites and see what kinds of book they publish. Do any of these look like yours?

If you know people who have written books with these publishers, you may also want to have a chat with them to see how things went. Were they good to work with, were they reliable and timely? Is there a lot of staff turnover? My own books have been mostly with Sage, and I have to say that they have been brilliant to work with. Not just professional; but supportive, friendly, and always encouraging. And they have the best parties (in fact, a colleague of mine recently wanted to publish with Sage just so that she could get invited!). I’m also very fond of PCCS Books and would definitely recommend them as a publisher to consider approaching. They’re a lot smaller than Sage, but have a real dedication to the books that they publish and care about the counselling and psychotherapy field very deeply. That makes a real difference when you feel like you are writing with a publisher that cares about the field—not just in it for the money.  

Write the proposal

Then you need to write a proposal. This is, perhaps, 10 pages or so, in which you describe what the book is, who it is for, a synopsis of chapters, and a CV, etc. I remember writing my first book proposal back in about 1988, and the mum of one of my friends, who had published with Penguin, made the point (very nicely) that if the writing in the proposal was that bad, how were the publishers ever going to think I could write a good book! So spend some time crafting the proposal and showing, straight away, that you can write.

Importantly, a lot of publishers will have their own format that they want proposals in. For instance, check out the Sage guidelines here. Even if you don’t want to publish with Sage, that will give you some great ideas about the kinds of things you need to cover in your proposal.

Generally publishers will want to see some examples of your writings. Again, send in something that reflects the kind of thing you want to write in the book. if you don’t have that yet, you may want to spend some time developing it before you write your proposal—just so you can show to yourself, as well as the publisher, that you can and do want to write in that way.  An example chapter or two can be a great thing to show to the publisher that the book can really work.

Of course, you could always write the book first and then send the whole manuscript to a publisher, but that's not always appreciated by publishers and can lead to a lot of wasted effort. Usually, publishers want to be involved in the development of a book, and will have a lot of good ideas about how to orientate it to their market.

Do I need an agent?

In this field, almost certainly not. If you think you’ve got a brilliant idea for a best selling ‘pop psych’ book, say for Penguin, then you may want to find a literary agent (you can search on the internet), but the amount of money in psychotherapy and counselling books means that it’s generally not worth it. And, yup, that’s right, not much money. So if you’re thinking that writing books in counselling and psychotherapy is going to make you your fortune, you’ll need to look elsewhere!

What next?

If the publishers think your proposal may be of interest to them, what they’ll then do is to send it out to some reviewers to see what they think, and to get feedback. You normally hear back in a few months. It’s not unusual to get rejected, particularly if you haven’t written before, and, of course, the thing is not to get demoralised but to learn from the feedback and see how you can revise your proposal for the next publisher.

If the publishers do want to take the book on, they’ll then send you out a contract to sign with various financial and timescale agreements. These are normally pretty straightforward, but a key thing to check is the royalties—that is, how much of the book sales you actually get. Normally, this starts off around 7%, so if a book sells for around £25, you’ll get about £2 per book. If you haven’t seen or signed one of these contracts before, try and see if there’s someone you know who has who can have a quick look over it and just check that it all looks OK.

Do I have to have an established publisher?

Absolutely not. There’s many ways to publish a book now that don’t involve going through the traditional route. It’s very easy, for instance, to do some self-formatting and then publish the book on your own website. Or just write the book as a series of blog posts. And that could be a good way of building up to a publication through more established routes over time. For instance, with my latest book on Integrating counselling and psychotherapy, I’d started off just writing a 20,000 word monograph to get the ideas out, and I put them on the internet. It was only several years later that I came back to this and fleshed it out into a full, 110,000 word text.

Is it all worth it?

Hm… 

 I couldn’t say it any other way than, for me, it’s an absolute bastard writing books. It’s a massive amount of work, commitment, focus, struggle—intellectually and emotionally. There’s time when I’ve felt completely out at sea, out of my depth, drowning. I’ve hated the book, hated myself for thinking I could write it, hated the whole process of sitting down for hours a day and trying to scratch out something of a meaning. But when you get that book finally in your hands, or when people say to you things like, ‘Ah, that book you wrote really helped me,’ or, ‘it made such a difference to my work,’ it does feel incredibly rewarding. Personally, I feel like, if I hadn’t of written, my life would feel so much more impoverished: I had so many things I wanted to say, and having that out there, in the public domain, forever, feels an amazing privilege. And it does make me want to say more things, to write more, to continue and deepen that dialogue with the world. So, yes, I guess, definitely worth it. Absolutely. But that’s just for me. And working out whether, for you, the pros are really worth the cons is, perhaps, the first step in the whole process.

 Very best of luck with it.

Presentations: Some Pointers

Present. Why? Because it’s a great way of getting your work out there: letting people know what you are doing, opening up conversation, getting feedback. When you present, you enter into dialogues with your community: people who can help you, encourage you, give you new ideas.

It’s scary. I know. I used to be absolutely phobic about presenting. I used to think, ‘What happens if I just clam up in front of all these people. Just stand there, dumbstruck, with all those eyes on me. Nowhere to go.’ But I did, really, push myself to present: to go for opportunities even if I knew I’d be terrified. And over time (albeit more time than I would have liked), it began to get easier.

For counselling and psychotherapy researchers, a great place to present your work is the annual research conference of the British Association for Counselling and Psychotherapy. It’s low key, friendly, and audiences are always really encouraging of people’s work. For counselling psychologists, another great opportunity to present is the BPS Counselling Psychology annual divisional conference.

Normally when you do a presentation, there will be a ‘Chair’ who will make sure you start and end on time, and possibly introduce you.

Research presentation are normally around 20 minutes, and then around 10 minutes for questions and discussion. But each conference will have their own guidelines.

Generally, conference delegates can pick and choose what they go to, and there’s likely to be a few strands of presentations running at once. So it hard to predict how many people will come to your talk, but it’s likely to be somewhere between about 10 and 50.

Research papers can either be presented individually, or as part of a ‘symposium’ (sometimes called a ‘panel’), where papers on a similar theme are grouped together. Normally, you can either submit as an individual paper or as part of a symposium—but if you do have colleagues doing similar work, creating a symposium can make for a more coherent set of presentations.

Prepare… prepare… prepare…

  • Know your timing: check that the length of your presentation fits into the allocated time slot. Be particularly wary of having much too much material for the time available.  Keep an eye on the time during your presentation and, if helpful, write on your notes where you should be up to by particular points, so you know if you need to speed up/slow down.

  • Practice your slides to get a good feel for them, and so you know what’s coming next.

  • Turn up to the room early and check your slide show is uploaded and works. Know the pointer, how to change slides, etc. Technological issues are often the biggest saboteur of a good presentation.

  • Try to introduce yourself to the Chair before you start (if there is one), and check how they are going to run things (in particular, how/whether they will let you know how much time you have left).

  • If you get anxious doing talks, think about how you could manage that. For instance, do you need things written out in detail to fall back on, or have breathing techniques ready if you get panicky?

  • Presenter View on Powerpoint can be a really helpful tool for being on top of your presentation. Essentially, it means that, when you present, you (and only you) can see what slides are coming up, and also any notes on your slides. It can be a bit technologically fiddly though.

  • Presenter View or not, it’s generally best to take along a printed off copy of your slides (say, 3 slides to a page), so that you can always quickly check content on other slides when you are doing presentation, and just in case the technology breaks down.

A great short video on what happens when you fail to prepare a presentation, and everything else you can do wrong, can be found here.

Slides

  • Keep the lines of texts per slide to a minimum. Generally no more than 6-10 lines of text per slide. If you have more to say, do more slides, they don’t cost anything! (I do really mean this one: so many presentations I see have 20+ lines of texts per slide, making the slide pretty ineligible).

  • Related to the above, font size shouldn’t normally be less than 30 points, and definitely not less than 16-20 points.

  • Texts should be bullet points, rather than complete sentences (so don’t have full stops at the end of them). Your bullet points should capture the essence of what you want to say (which you can then expand upon verbally), rather than spelling the point out in full.

  • Try to avoid

    • Sub bullet points,

      • And sub-sub bullet points.

        • The slides start to get very messy.

  • Be consistent in your formatting: e.g. fonts, type of bullets, colour of headings.

  • If you have text on your slides, talk ‘to’ it. Don’t have text on the slide that you never refer to. (Though it’s OK to say things that aren’t on your slide).

  • Use the space on the slides—make text large rather than small text squashed away.

  • You don’t need line spacing between your bullet points. If you take those out you can make your font larger.

  • Try to avoid too many citations in your bullet points as they can be distracting. You can cite sources at the bottom of the page or have a page of references at the end of the presentation if people want to follow up. Having said that, if you’re discussing a key text, make sure you give a reference so that your audience can follow up.

  • Sans serif fonts (e.g., Arial, Tahoma, Century Gothic) are generally more suited to presentations than serif fonts (e.g., Times New Roman, Palatino). NCS: Never Comic Sans!

  • Try to use images/graphics wherever possible, ideally each slide.  You can also embed videos (but check sound works before your presentation). Images and videos can be a great way of conveying the reality of your research: for instance, a photo of the room where the interview took place, or a short video of you doing the coding (bear in mind confidentiality, of course).

  • Diagrams can be really helpful, but do make sure you spend time talking them through and explaining what different elements mean. Don’t just leave it up to your audience to work it out for themselves.

  • Don’t make slides too complex/‘flashy’: for instance, by using transition sounds.  Everyone hates transition sounds!

  • Having said that, a simple transition between slides, like ‘Fade’, can be a nice way of going from slide to slide.

  • ‘Animations’ allow you to present one bullet point at a time, and can be helpful for ensuring that you and your audience are on the same points. Again, though, just use simple entrance animations, like Fade, so that it doesn’t detract from your content.

  • For a research presentation, it’s generally fine to use the standard sections of a research paper to structure your talk: Introduction (including literature review), Aims, Methods, Results, Discussion. Headings can be on separate slides to keep the sections really clear.

  • Give clear titles to each slide so that the audience know what you are trying to say.

  • Don’t scrimp on presenting your results: they’re often the most important and interesting part of your paper, so ensure you leave a proper amount of time to talk through them (say 50% of your overall time, if a qualitative presentation).

  • Everyone uses Powerpoint—think about trying Prezi.

  • Watch copyright—you shouldn’t use images that aren’t in the public domain. You can find many images that are available for reuse via Google Search/Images/Settings/Labelled for reuse.

 ConnectING with your audience

  • This is the key to everything: talk to your audience. Try to connect with them. Imagine, for instance, that they are a friend that you really want to explain something to. You’re not trying to be smart, or clever, or get them to approve of you—you just want to explain something to them about what you’ve done, what you’ve found, and what it means. So breathe, focus, speak to the people in the room (or online). Try not to just rattle through what you have to say.

  • That means trying, yourself, to connect with the ‘story’ of what you are saying: if it’s meaningful to you it’s more likely to be meaningful to your audience.

  • Remember that, nearly always, your audience are there to learn from you—not to judge you. They haven’t come along to your presentation thinking, ‘Hmm, I wonder if [insert your name here] is a good presenter or not. I’d really like to know.’ In fact, the harsh truth is that they’re almost certainly not thinking much about you at all. Rather, they’re thinking, ‘Hmm… I’d be interested to know more about [insert your topic here]’. So the question you should be asking yourself is not ‘How can I prove I’m good enough?’ but, ‘What can I teach these people?’

  • Lead your audience through your talk. You may be really familiar with your material, but they are unlikely to be. So explain things properly: from why you did your research, to what your findings mean, to what it says, ultimately, about clinical practice.

  • Know who your audience is and adjust accordingly. For instance, a group of experienced practitioners may know, and want, very different things from a group of early stage researchers. Think about what your audience will want from the talk. And what they might already know (that you don’t need to repeat)?

  • Try not to read directly from your notes, or from your slides. Best to use them as stimuli.

  • Avoid jargon or lots of acronyms. Keep it as clear and easy to understand as you can. If you need to use acronyms, explain clearly what they mean.

  • Speak loud and clear—check people can hear you, if need be, particularly at the back.

  • Watch that you’re not talking too fast, particularly if you’re anxious. Try the talk out with a friend/colleague and get some honest feedback from them.

  • Pace your talk, so that you have enough time for all of it. It’s a classic mistake to get very caught up in the first part of your talk, and then have to rush the rest (and often the most interesting bits).

  • Make sure you leave time for questions—so that you’re audience can really engage with you.

  • Don’t be defensive if asked questions: accept that there may be things to develop in your paper if you can see that.

  • It’s really bad form to run over time, as it means you’re eating into the next person’s allocated slot (or everyone’s coffee/lunch). So if you’re asked to stop, stop. (I once saw a presentation where the speaker, already running 30 minutes over time, starting asking the audience whether or not they thought he should be hauled off. Very, very awkward!)

  • It’s fine to bring yourself in to the presentation, and often that’s a way of helping the audience connect with you. For instance, why did you, personally, want to do this study? What did you, personally, get out of it?

  • Humour can be a great way of connecting, and cartoons can often lighten a talk and engage and audience. But don’t force humour if it’s not there or if it’s not ‘you’.

  • And, finally, don’t stand in front of the projector!

Posters

If you’re not keen on presenting a paper orally, you can always present a ‘poster’. That can be particularly appropriate if your work is still in progress. And it’s another great work of initiating dialogue around your work with other members of the counselling community.

DISCLAIMER

 The information, materials, opinions or other content (collectively Content) contained in this blog have been prepared for general information purposes. Whilst I’ve endeavoured to ensure the Content is current and accurate, the Content in this blog is not intended to constitute professional advice and should not be relied on or treated as a substitute for specific advice relevant to particular circumstances. That means that I am not responsible for, nor will be liable for any losses incurred as a result of anyone relying on the Content contained in this blog, on this website, or any external internet sites referenced in or linked in this blog.

The Discussion Section: Some Pointers

The following blog is for Master’s or doctoral level students writing research dissertations in the psychological therapies fields. The pointers are only recommendations—different trainers, supervisors, and examiners may see things very differently.

The aim of a discussion section is to discuss what your findings mean, in the context of the wider field.

As with all other parts of your dissertation, make sure that your Discussion is actually discussing the question(s) that you set out to ask.

It’s really important that your Discussion doesn’t just re-state your findings (aside from a brief summary at the start). It’s often tempting to reiterate results (just in case the reader didn’t get them the first time!), but now’s the time to move on from your findings, per se. Structuring your Discussion in a different way from your Results can be a good way of trying to ensure this. So, for instance, if you’ve presented your Results by theme, you might want to structure your Discussion by stakeholder group or by research questions.

Generally, you shouldn’t be presenting raw data in your Discussion: for instance, quotes or statistical analyses. That goes in your Results.

Similarly, try to avoid referencing lots of new literature in your Discussion. If it’s so relevant, it should be there in your Literature Review.

Make sure that your Discussion does, indeed, discuss your findings. It shouldn’t just be the second half of your Literature Review: something which bypasses your own research. Emphasise the unique contribution that your findings make, and focus on what they contribute to knowledge. Be confident and don’t underplay the importance of your own findings.

At the same time, don’t over-state the implications of your findings (particularly with regard to practice). Be realistic about what they mean/indicate, in the context of the limitations of your study, as well as its strengths.

This is your chance to be creative, exploratory, and to investigate specific areas in more detail, but try to ensure that it’s always grounded in the data: what you found or what others have found previously. So not just wild speculation.

What’s unexpected in your results? What’s surprising? What’s counter-intuitive? What’s anomalous? Your Discussion is a great opportunity to bring these out to the fore more fully and explore them in depth.

Typical sections of a discussion section (often in approximately this order)

  • Brief summary of your findings (but keep it brief—just a concise but comprehensive paragraph or two).

  • What your findings mean, in the context of the previous literature. So, for instance, how they compare with/contrast/confirm/challenge previous evidence and theory. This is also an opportunity for you to untangle, and to try and explain, complex/ambiguous/unexpected findings in more depth.

    • This would normally be the bulk of your Discussion. It may be appropriate to structure this section by your research questions, or by the themes in your results. If you do the latter, though, as above, be careful that you’re not just reiterating your findings.

    • Remember that you don’t need to give equal weight/space to all your findings. If some are much more interesting/important than others, it’s fine to focus your Discussion more on those; though all key findings should be touched on at some point in the Discussion.

  • Limitations. This should be a good few paragraphs. Try to say how the limitations might have affected the results (e.g., ‘a volunteer sample means that they may have been more positive than is representative’) rather than just what the flaws in the study were, per se.

    • Be critical of what you did; but from a place of reflective, appreciative awareness, rather than self-flagellation. The point here is not to beat yourself up, but to show that you can learn, intelligently; just as you did something, intelligently.

  • Implications for clinical practice. Also, if relevant, implications for policy, training, supervision, etc.

    • Try to keep this really concrete: what would someone do differently, based on what you found.  So, for instance, not just, ‘These findings may inform practitioners that….’ But, ‘Based on these findings, practitioners should….’

  • Specific implications for your specific discipline: e.g., counselling psychology/counselling/psychotherapy.

  • Suggestions for further research.

  • Reflexivity: what have you learnt from the study, both in content and in practice.

Conclusion: this can be a brief statement bringing all your thesis together.

Appendices

Following your references, you are likely to want to append various documents to your thesis. These can include:

  • Participant-facing forms: e.g., information sheets, consent forms, adverts.

  • Full interview schedule.

  • Additional quantitative analyses and tables.

  • A transcript of one interview (but bear in mind confidentiality—this may not be appropriate). This could also show your coding of that interview.

  • All text coded under one particular theme/subtheme, for the reader to get a sense of how you grouped data together (again, bear in mind confidentiality).

(Image by Muhammad Rafizeldi, Creative Commons Attribution-Share Alike 3.0 Unported license)

The Results Section: Some pointers

The Results section is the beating heart of your research. You’ve set out on your quest. You’ve told the reader what’s known so far (the Literature Review), and then how you’re going to answer your question (the Methods). Now, finally, after thousands of words of wait, you’re going to tell us what you’ve found out. How did clients, for instance, actually experience transactional analysis, or was there any relationship between therapists’ levels of self-awareness and their outcomes? So fanfare, please, because it’s what we’re all dying to find out about. Put more prosaically, don’t just mutter away your finding under tables or jargon or lengthy quotes that never really tell us what you actually discovered. Own it, make it exciting, tell us—as clearly and succinctly as you can—about the answers that you’ve found.

Two thing, I think, can have a tendency to act as killjoys to the Results section. First is social constructionism. I say this as a (sort of) social constructionist myself, but the problem is that this mindset can take you so far away from the idea that there is anything out there ‘to be discovered’ that the findings, themselves, become almost an irrelevancy. Instead, the focus becomes on the method and the epistemological positioning behind it; and while that might be of some interest, personally, I think it’s only a vehicle to what is most exciting and interesting about research, and that’s discovery. Second, is the researcher’s own lack of self-confidence. If you are a novice researcher, you may feel that what you, yourself, are discovering isn’t really worth much, so you don’t feel there’s really much point emphasising it. If, at some level, you do feel that, it’s worth reflecting on it and thinking about what your research really can contribute. You need to feel, and you need to show, that you are adding something, somewhere.

Another general point: by the time you completed data collection, your data—whether it’s qualitative or quantitative—is likely to be large, complex, messy, and easily overwhelming. Like a dense forest. And that means that when you write it up, your reader—let alone yourself—can easily get lost. So a good write-up really needs to guide the reader through the results. Make it easier for them to find their way—not harder. Remember that you will have spent weeks, maybe months, getting to know your data, so what might seem obvious and clear to you may be entirely unfamiliar to your reader.  Hold their hand as you walk them through it. And if there’s things that, actually, you don’t really understand, don’t just present it to your readers in the vague hope that they’ll get it even if you don’t. Remember, you’re the expert, the leader here (see The Research Mindset). So you need to process and digest the data, make full sense of it yourself, and then present it to your readers in a way that they can easily grasp. Think ‘bird digesting food before it feeds it to its chicks’. You need to do the work of digestion, so that what the reader is fed is as consumable and nourishing as possible.

When leading your reader through the forest of your findings, one really important thing is to try and be as consistent as possible in how you report your results. For instance, don’t report frequencies in the first half of your qualitative write-up but not in the second; or shift from two to three decimal places after the first few analyses. Make rules for yourself about how you are going to report things (and write them down, if necessary) and then stick to them all the way through. And keep the same terms throughout. If, for instance, you switch between ‘patients’ and ‘participants’ and ‘young people ‘ to refer to the people who took part in your study, your reader might be wondering if these are all the same things or different. And, particularly importantly, use exactly the same terms for themes, categories etc. throughout. It might be obvious to you, for instance, that ‘Boosting Self-Esteem’ is the same as ‘Building Self-Confidence’, but for the reader who isn’t inside your head it can get really confusing trying to work out what is what if the terms keep changing.

Qualitative analysis

For a 25,000 word thesis, a qualitative Results section may be 8,000 words or so.

That means it is can be a good idea to give a table of the overall structure of your analysis and themes/subthemes at the start of your Results. However, if you give a table, you should ensure that the wording of the themes/subthemes on the table matches, exactly, the headings/subheadings in your narrative account of the results. Otherwise, it can confuse them even more!

Frequency counts in the table and/or in the text (usually the number of participants who were coded within a particular theme/subtheme), can help give the reader a sense of how representative different themes/subthemes are.  Some researchers dislike this as it can feel too ‘quanty’ (‘small q’) and inconsistent with a ‘Big Q’ qualitative worldview (for discussion of big and small q qualitative research see, for instance, here). It may also be seen as suggesting more precision and generalisability than there actually is.  One option, in the narrative, is to use a system that labels different frequencies within broad bands. The most common one was developed in consensual qualitative research (see, for instance, here), and uses the terms:

  • ‘general’: for themes that apply to all cases

  • ‘typical’: for themes that apply to at least half of cases

  • ‘variant’: for themes that apply to at least two or three, but fewer than half, of cases

An alternative ‘scoring scheme’ for qualitative analysis is detailed here.

In your narrative, it’s generally a good idea to use subheadings (and, if necessary, sub-subheadings) to break the analysis up, and to make it clear to the reader where they are in the account. Nearly always, these would be a direct match to your themes/subthemes/sub-subthemes. Alternatively, for your sub-subthemes, you can italicise the title in the text (making sure it matches what is in the table) to help orientate the reader.

Direct quotes from your participants are an important way of evidencing your themes and subthemes, and really bringing your analysis ‘to life’. They make it clear that your analysis is not just based on theoretical conjectures, but on the realities of people’s narratives and experiences.

However, make sure that you integrate/summarise, in your own words, what participants are saying, rather than just presenting long series of quotes with just a few words in between. Anyone can cut and paste quotes from a transcript to a dissertation. If that’s all your doing, it may fill up your word count, but it really doesn’t show your understanding of what your participants are saying, and how their different accounts fit together. So don’t use quotes as a substitute for a comprehensive and thorough analysis of what your data mean.  And where you do quote your participants, always make it clear what you are trying to ‘say’ with that quote (rather than just dropping it in, and leaving the reader to work it out for themselves), for instance:

  • ‘Sarah’s experiences at the start of transactional analysis illustrate how this approach can be experienced as very holding: “When I first went to the therapist…”

  • ‘Some participants said that they really valued the psychoeducational component of transactional analysis: ‘I immediately recognised my Parent, Adult, and Child ego states, and found it could help me make sense of so many of my problems’ (Ashok, Line 234).

  • Although most participants like the psychoeducational aspect of transactional analysis, a couple had mixed responses. Gemma, for instance, said:

She kept on going on about ‘strokes’, and I just- it seemed a bit jargony…

Along these lines, while long quotes can be very helpful in giving the reader an extended sense of what participants have said, if they illustrate, or evidence, many different points, you may be better off breaking them down into shorter segments so you can clearly explain what each part means.

The format of text in your results can be the same as throughout the rest of your thesis. So, for instance, only indent quotes that are 40 words or more long, don’t italicise quotes, put full stop before the reference for the quote if indented (and after if in the body of the text).

For referencing quotes, you should normally give the pseudonym of the person saying it, and a reference to where it is in their transcript (e.g., line number). So, for instance, ‘… (Mary, Line 230)’. 

Normally, references to other literature should not be in the Results. Save that for the Discussion.

Finally, above and beyond all the pointers above, it’s important that the way you write your results is consistent with your method and epistemology.  So, for instance, if you have adopted a social constructionist epistemology, don’t start making realist claims like, ‘Men were more defensive than females…’   Generally, the more realist your approach, the more you may want to use tables, frequency counts, etc.; while more constructionist epistemologies may lead to less structured and quantified analyses.   

Quantitative analysis

A quantitative Results section is likely to be shorter than a qualitative one, so for a 25,000 word thesis, perhaps 4,000 words or so, though this can vary enormously depending on content.

Rather than just presenting stats and leaving it to the reader to interpret it, make sure you explicitly state what your findings mean (e.g., ‘Chi-squared tests indicate that men were significantly more likely than women to…’). In particular, be clear about which group was higher/lower than which.

In describing your findings, use precise language. Is it ‘significant’/’non-significant’?, refer to the specific effect size and stats: not, ‘This seems to indicate that men were a bit more empathic than women,’ but ‘Men were significantly more empathic than women (F = …).

Remember that, if you are using inferential tests, something is either significant or not. You can generally get away with talking about a ‘trend’ if the p value is between .1 and .05, but be very cautious; and make sure you don’t spend a lot of time interpreting or discussing non-significant findings.

Don’t just rely on significance tests. Give confidence intervals wherever possible and also effect sizes.

Be consistent in how many decimal points you use, and use only as many as is meaningful.  Does it really help the reader, for instance, to know results down to four decimal points? Often just one is enough for means and standard deviations, maybe two or three for p-values. That can also make it clearer for the reader to see what the findings are.

Remember that, with the vast majority of statistical tests, you cannot prove the null hypothesis, so be sure to avoid phrases like: ‘This indicates that men and women had equivalent levels of empathy,’ rather, ‘the difference in levels of empathy between men and women was non-significant.’

Although graphs can look pretty (especially with lots of colours), tables are often a more precise means of presenting data, and generally mean that you can present much more data at once.

It’s rarely a good idea to just cut-and-paste SPSS tables – better to re-enter the data as a Word table so that you can get the formatting of the table appropriate to the journal.

The APA Publication Manual (7th edition) has some great guidance on how to format and present all aspects of quantitative statistics.  It can also help you make sure that you stay consistent in how you format your Results—as well as other parts of your paper. An essential companion, particularly if you are doing quantitative analysis. Further pointers on quantitative analysis are available here.

Disclaimer

 The information, materials, opinions or other content (collectively Content) contained in this blog have been prepared for general information purposes. Whilst I have endeavoured to ensure the Content is current and accurate, the Content in this blog is not intended to constitute professional advice and should not be relied on or treated as a substitute for specific advice relevant to particular circumstances. That means that I am not responsible for, nor will be liable for any losses incurred as a result of anyone relying on the Content contained in this blog, on this website, or any external internet sites referenced in or linked in this blog.

The Methods Section: Some Pointers

The following blog is for Master’s or doctoral level students writing research dissertations in the psychological therapies fields. The pointers are only recommendations—different trainers, supervisors, and examiners may see things very differently.

What should go into the Methods chapter of a thesis, and how much should you write in each area? The headings, below, describe the typical sections, content areas, and approximate length . The suggested word lengths are in the context of a 25,000-30,000 word thesis, and may be a bit more expanded for a longer dissertation (and obviously more condensed for a shorter one).

Epistemology

(Approx. 2,000-3,000 words).

This is often a requirement of Master’s or doctoral level theses, and is a key place in which you can demonstrate the depth and complexity of your understanding. This may be a separate chapter on its own, or placed somewhere else in the thesis.

  • Critical discussion of epistemology adopted (e.g., realist, social constructionist)

  • Links to actual method used

  • Consideration/rejection of alternative epistemologies. 

Design

(Approx. 50-500 words).

  • Formal/technical statement of the design: e.g., ‘this is a thematic analysis study drawing on semi-structured interviews, based in a critical realist epistemology’

  • Any critical/controversial/unusual design issues that need discussing/justifying.

Participants

(Approx. 500 words).

  • Site of recruitment: Where they came from/context

  • Eligibility criteria: inclusion and exclusion

  • Demographics (a table here is generally a good idea: can by one participant per row if small N, or one variable per row if large N)

    • Gender

    • Age (range/mean)

    • Ethnicity

    • Disability

    • Socioeconomic status/level of education

    • Professional background/experience: training, years of practice, type of employment, orientation

  • Participant flow chart/description of numbers through recruitment: e.g., numbers contacted, number screened, numbers consented/didn’t consent (and reasons). Also organisations contacted, recruited, etc.  

Measures/Tools

(Approx. 500 words).

  • Interview schedule

    • Nature of interviews: e.g., structured/semi-structured? How many questions?

    • Give key questions

    • Prompts?

    • (Full schedule can go in appendix)

  • Measures (including any demographics questionnaire): a paragraph or two on each

    • Brief description

    • Background

    • What it is intended to measure

    • Example item(s)

    • Psychometrics:

      • reliability (esp. internal reliability, test/retest)

      • validity (esp. convergent validity)

Procedure

(Approx. 500-1000 words).

  • What was the participants’ journey through the study: e.g., recruitment, screening, information about the study, consent, interview (how long?), debrief, follow up

  • Nature of any intervention: type of intervention (including manualisation, adherence, etc), practitioners…  

Ethics

(Approx. 500 words).

  • Statement/description of formal ethical approval

  • Key ethical issues that arose and how they were dealt with 

Analysis

(Approx. 1,000-2,000 words).

  • What method used

  • Critical description of method (with contemporary references)

  • Rationale for adopting method

  • Consideration/rejection of alternative methods

  • Stages of method as actually conducted (including auditing/review stages) 

Reflexive statement

(Approx. 250 words).

Remember that the point of your reflexive statement here is not to give a short run-down of your life. It’s about disclosing any biases or assumptions you might have regarding your research question. We will all have biases, and by being open about them you can be transparent in your thesis and all the reader, themselves, to judge whether your results might be skewed in any way.

  • What’s your position in relation to this study?

  • What might your biases/assumptions be?