The HE Green Paper: (Don’t) Read it and Weep – Part 1: The TEF & Social Mobility

The Disorder Of Things

Britain’s Conservative government recently released its much-awaited (or much-dreaded) ‘green paper’ on higher education (HE), a consultation document that sets out broad ideas for the sector’s future. Masochistically, I have read this document – so you don’t have to. This first post describes and evaluates the centrepiece of the green paper, the Teaching Excellence Framework (TEF), and measures on ‘social mobility’.

View original post 3,063 more words


The HE Green Paper: (Don’t) Read it and Weep – Part 2: Completing the Market

The Disorder Of Things

This post continues where Part 1 left off.

The real goal of the green paper is to accelerate the formation of a fully functioning market in HE – as has already been discussed elsewhere by the brilliant Andrew McGettigan. The opening move was HEFCE’s QA consultation earlier this year which, as I explained on TDOT, was an attempt to dilute quality standards to make it easier for ‘alternative’ (i.e. private) providers to enter the market. Whereas HEFCE hid behind technocratic jargon, however, the green paper openly announces the government’s ‘clear priority’ to ‘widen the range’ of HE providers (p.50). ‘Our aspiration is to remove all unnecessary barriers to entry’ and create a ‘level playing field’ (p.42).

View original post 1,088 more words

How research-informed practice stood up to the pseudo-science of inspection: defending an ungraded approach to the evaluation of teachers


This post tells the story of a university partnership of teacher educators’ experience of an Ofsted inspection of its Initial Teacher Education (ITE) provision in March 2013. Building on a position paper that was written at the time of the inspection, this post outlines how we defended our position on not grading our student teachers and shares some of the underpinning principles of our philosophy. Given the recent shift in Ofsted policy to remove the grading of individual lesson observations from school inspections, this post is very timely as it discusses some of the challenges faced by a department that has not only never used the Ofsted 4-point scale to assess its student teachers during observations, but resisted the use of numerical grading scales across its programmes as a whole.

Few areas of practice have caused as much debate and unrest amongst teachers in recent years as that of lesson observation, particularly graded observations and the way in which they have been used as summative assessments to rank teachers’ classroom performance against the Ofsted 4-point scale. Recent research in the field has described how graded lesson observations have become normalised, highlighting Ofsted’s hegemonic influence and control over education policy and practice (e.g. O’Leary 2013). At the same time, they have been critiqued for embodying a pseudo-scientific approach to measuring performance, as well as giving rise to a range of counterproductive consequences that ultimately militate against professional learning and teacher improvement (e.g. O’Leary and Gewessler 2014; UCU 2013). 


Unlike the vast majority of other university ITE providers in England, the post-compulsory education (PCE) department at the University of Wolverhampton has never used graded observations on its programmes. The underpinning rationale for adopting an ungraded approach to the assessment of our student teachers did not emerge arbitrarily but was developed collaboratively over a sustained period of time. This approach was underpinned by a core set of principles and shared understandings about the purpose and value of our ITE programmes, as well as being informed by empirical research into the use and impact of lesson observations in the Further Education (FE) sector and on-going discussions with our partners and student teachers. Given that our approach went against the grain of normalised models of observation, we knew that our programmes would be subject to heightened scrutiny and interrogation by Ofsted when it was announced that all the university’s ITE programmes would be inspected in March 2013.

The tone was set soon after the arrival of the inspection team on the first day when the lead inspector asked the PCE management team to rate the quality of its provision against Ofsted’s 4-point scale. This was despite the fact that the team had chosen not to apply this grading scale in its self-evaluation document (SED), which all providers were required to complete and submit at the end of each year and to which Ofsted had access before the inspection. But why did the partnership adopt this stance? It is important to emphasise that our resistance to embracing Ofsted’s ‘dominant discourses’ (Foucault 1980) and normalised practice was not based on any wilful refusal to comply or obey their authority as the regulators of quality for ITE provision, but driven by more fundamental concerns regarding the legitimacy and reliability of its assessment framework and the impact of that on teachers in training. Needless to say this epistemological positioning did not sit easily with the inspection team as it presented them with certain challenges that they were unaccustomed to, some of which are discussed further below.

Evaluating performance

It was a strongly held view across our partnership that the use of a metrics-based approach was neither the most appropriate nor the most effective means of fostering our student teachers’ development, nor indeed of measuring the level of performance required to meet the ‘pass’ threshold criteria of our programmes. Our partnership staff comprised largely experienced teacher educators who were comfortable and confident of being able to make judgements about the progress and performance of their students against the pass/fail assessment framework used on the programmes. In some ways this was akin to the notion of ‘fitness to practise’ used by other professions such as health. This ‘fitness to practise’ was initially mapped against the professional standards in use at the time in the FE sector (LLUK 2006) and more recently against the Education and Training Foundation’s (ETF) revised standards (ETF 2014). As the PCE partnership had been actively engaged with these standards through year on year collaborative work to revise and refine their application to its ITE programmes, there was a shared ownership of the assessment by those working on the programme. In contrast, we were not convinced that the Ofsted 4-point scale could be applied with the same rigour, reliability and appropriateness to assess students’ attainment as our existing assessment framework and criteria, whereby students were either judged to have satisfied the criteria or not. In other words, whilst all those teacher educators working on the programmes were clear as to what constituted a pass/fail and were confident in applying these criteria accurately and consistently, the same could not be said about the interpretation and application of Ofsted’s 4-point scale.

In their study into the grading of student teachers on teaching practice placements in Scotland, Cope et al (2003: 682) found that the success of such practice depended on ‘a clearly reliable and valid system of assessment of the practice of teaching’ and concluded that ‘the evidence available suggests that this does not currently exist’. This is not a phenomenon specific to observation as a method of assessment, but reflects widely held beliefs among key researchers in the field of assessment such as Gipps (1994: 167), who argued back in the 1990s that ‘assessment is not an exact science and we must stop presenting it as such.’ The danger, of course, is that the inherent limitations of practice such as numerically grading performance are often overlooked and the resulting judgments are given far more weight and authority than they can realistically claim to have or indeed deserve.

Prioritising teacher development

Our ITE programmes are built on a developmental philosophy in which the student teacher’s growth is prioritised. Staff working on the programmes are committed to helping their students to develop their pedagogic skills and subject knowledge base. It was therefore their belief that judging them against a performative, numerical grading scale of 1-4 would compromise that commitment and jeopardise the supportive focus of the teacher educator and mentor’s relationship with their students. The partnership also benefitted from being involved in and discussing the latest research into lesson observation as one of the university members of staff specialised in this particular area.

As mentioned above, recent research into the use of graded observation in FE reveals how it has become normalised as a performative tool of managerialist systems fixated with attempting to measure teacher performance rather than actually improving it (e.g. O’Leary 2012). The teacher educators and mentors in the PCE partnership saw their primary responsibility as that of helping to nurture their student teachers as effective practitioners rather than having to rank their performance according to a series of judgemental labels (i.e. ‘outstanding’, ‘inadequate’ etc.) that were principally designed to satisfy the needs of external agencies such as Ofsted within the marketised FE landscape and carried with them absolutist judgements that were inappropriate to their isolated, episodic nature. This emphasis on measuring teacher performance was also seen as responsible for what Ball (2003) refers to as ‘inauthenticity’ in teacher behaviour and classroom performance during assessed observations. This is typically manifested in the delivery of the rehearsed or showcase lesson as the high stakes nature of such observations results in a reluctance to want to take risks for fear of being given a low grade. Teachers are thus aware of the need to ‘play the game’, which can result in them following a collective template of good practice during observation. Yet being prepared to experiment with new ways of doing things in the classroom and taking risks in one’s teaching is widely acknowledged as an important constituent of the development of both the novice and experienced teacher.

Furthermore, findings from two separate studies on observation in FE (e.g. O’Leary 2011; UCU 2013) have revealed some of the distorting and counterproductive consequences of grading on in-service teachers’ identity and professionalism. Staff in the PCE partnership, many of whom are FE teachers themselves, were determined to protect their student teachers from such consequences during their time on the programme. This did not mean, however, that they avoided discussing the practice of grading teacher performance with them or confronting some of the challenging themes and issues associated with it. On the contrary, this was a topic that was addressed explicitly through professional development modules and wider discussions about assessment and professionalism as part of the on-going critically reflective dialogues that occurred between teacher educators, mentors and students throughout the programme.

Developing critically reflective teachers

The university’s PCE ITE programmes are underpinned by the notion of critical reflection. Brookfield (1995) argues that what makes critically reflective teaching ‘critical’ is an understanding of the concept of power in a wider socio-educational context and recognition of the hegemonic assumptions that influence and shape a teacher’s practices. The PCE partnership viewed the use of graded observations as an example of one such hegemonic assumption. Thus the perceived or intended outcomes of graded observations (i.e. improving the quality of teaching and learning, promoting a culture of continuous improvement amongst staff etc.) were not always the actual outcomes as experienced by those involved in the observation process. And then, of course, there was the thorny issue of measurement.

The ongoing fixation with attempting to measure teacher performance is symptomatic of a wider neoliberal obsession of trying to quantify and measure all forms of human activity, epitomised in the oft-quoted saying that ‘you can’t manage what you can’t measure’, a maxim that has its roots in a marketised approach to educational improvement and one which seems to shape Ofsted’s inspection framework. During the inspection, it became apparent that the PCE partnership’s ungraded approach was problematic for Ofsted. Although when I asked the lead inspector directly at a feedback meeting if the use of a grading scale was considered an essential feature of being able to measure teachers’ progress and attainment, he categorically stated that was NOT the case nor did Ofsted prescribe such policy, he later contradicted this in his final report by maintaining that as the partnership did not grade, it was ‘difficult to measure student progress from year to year or the value that the training added in each cohort’. In spite of the presentation of interwoven sources of qualitative evidence (tutor/mentor/peer evaluations, self-evaluations, integrated action/development plans, critically reflective accounts etc) illustrating these student teachers’ journeys throughout their programmes of study, the inspection team was reluctant or even unable to conceptualise the notion of improvement unless the outcome was expressed in the form of a number. And why is that? Because, of course, reading such qualitative accounts are more time consuming and ‘messier’ than the reductive simplicity of allocating a number to something, however spurious that number might be. This reveals the extent to which ‘managerialist positivism’ (Smith and O’Leary 2013) has become an orthodoxy and Ofsted its agent of enforcement. Despite that, the partnership team defended its practice and emphasised how the broad range of evidence captured in the combination of formative and summative assessments provided a rich tapestry of these student teachers’ progress and attainment throughout the programme and ultimately one that was more meaningful than the allocation of a reductive number.


Ball, S. (2003) The teacher’s soul and the terrors of performativity, Journal of Education Policy, 18(2), pp. 215-228.

Brookfield, S. D. (1995) Becoming a Critically Reflective Teacher. San Francisco, CA: Jossey-Bass.  

Cope, P., Bruce, A., McNally, J. and Wilson, G. (2003) Grading the practice of teaching: an unholy union of incompatibles. Assessment & Evaluation in Higher Education, 28(6), pp. 673-684.

Education and Training Foundation (ETF) (2014) Professional Standards for Teachers and Trainers in Education and Training – England. Available at:

Foucault, M. (1980) Power/Knowledge – Selected Interviews and Other Writings 1972-1977. Brighton: The Harvester Press.

Gipps, C. (1994) Beyond Testing: Towards a Theory of Educational Assessment. London: Falmer Press.

Lifelong Learning UK (LLUK) (2006) New overarching professional standards for teachers, tutors and trainers in the lifelong learning sector. London: LLUK

O’Leary, M. (2011) The Role of Lesson Observation in Shaping Professional Identity, Learning and Development in Further Education Colleges in the West Midlands, unpublished PhD Thesis, University of Warwick, September 2011.

O’Leary, M. (2012) Exploring the role of lesson observation in the English education system: a review of methods, models and meanings. Professional Development in Education, 38(5), pp. 791-810.

O’Leary, M. (2013) Surveillance, performativity and normalised practice: the use and impact of graded lesson observations in Further Education Colleges. Journal of Further and Higher Education, 37(5), pp. 694-714.

O’Leary, M. & Gewessler, A. (2014) ‘Changing the culture: beyond graded lesson observations’. Adults Learning– Spring 2014, 25: 38-41. 

Smith, R. & O’Leary, M. (2013) New Public Management in an age of austerity: knowledge and experience in further education, Journal of Educational Administration and History, 45(3), pp. 244-266.

University and College Union (UCU) (2013) Developing a National Framework for the Effective Use of Lesson Observation in Further Education. Project report, November 2013. Available at:

Observation rubrics – a response to @joe_kirby

@joe_kirby’s recent post  makes reference to my book in the context of a wider discussion regarding the ongoing use of lesson observation in the English education system. As all readers will be aware, observation is a hot topic that continues to generate much debate across the profession, albeit often for the counterproductive consequences of its predominantly performative use. The fact that teachers like Joe and others have written numerous blogs about it recently reinforces the idea that it continues to provoke strong emotions across the education sector. In his post Joe selects a series of quotes/extracts from the book in an attempt to encapsulate some of the thematic discussion and the main arguments I present. It’s no mean feat trying to capture some of the key arguments and topics covered in the book’s nine chapters in a blog entry but Joe’s inclusion of the following summarising statement from the book towards the start does a good job of setting the tone:

The high-stakes nature of performance management-driven observation for monitoring and measuring militates against professional development; school leaders must challenge the hegemony of graded observations and redesign observation as a tool for reciprocal learning, decoupled from summative high-stakes grading

For those people who would like to read detailed reviews of the book, there is one here and another here. And, of course, there are shorter ones on Amazon too. Following on from Joe’s blog, I’d I’d just like to add a couple of points of clarification and extend the discussion further. This post is certainly not intended to be a long one, nor indeed going to repeat things I’ve written/spoken about before regarding observation. If anyone is interested in reading my previous work on the topic, you can access journal articles, reports, talks etc for FREE by clicking on my web page, where I regularly post my publications/output. Firstly, Joe makes a point about feeling ‘uneasy’ with the ‘prescription’ of the Ten principles of ‘good teaching’ graphic, which appears in Chapter 6 ‘Being an Effective Teacher – Models of Teacher Effectiveness’. I’m unsure if Joe has misread that particular extract but this is what it actually says in the book in relation to the ‘Ten principles’:

Despite the difficulties previously discussed in defining good teaching, this does not mean to say that it is impossible or pointless, as Moore suggested above, to devise a set of ‘guiding principles’. It is one thing to produce a prescriptive list of ‘dos’ and ‘don’ts’ as to what constitutes good teaching but another to theorize about some of its underpinning principles. With this in mind, what might such a set of principles or assumptions comprise?

Table 6.1 below is my attempt to produce a broad set of principles of ‘good teaching’, though not necessarily in a particular order. Table 6.1 is not meant to be an exhaustive list, but should provide a broad framework for discussing the topic within and across institutions. It might also be used as a set of prompts on which to base the development of a more tailored instrument for assessing practice across the institution through the medium of classroom observation and other relevant mechanisms. (O’Leary 2014a: pp. 97-98)


Thus this list of Ten Principles of ‘Good Teaching’ are simply a stimulus for debate and certainly not meant to be used prescriptively. Much of the discussion in that particular chapter of the book explores research into the notion of teacher effectiveness and makes it clear that it is a contested terrain with conflicting findings from a host of international studies. Joe contends that ‘any selective list of what makes good teaching will never be agreed upon’. One thing that is clear from past and current research into teacher effectiveness and attempts to define it is that it is indeed a thorny topic that divides opnions. However, that shouldn’t stop us from discussing it and developing our knowledge and understanding further of the qualities and attributes of the effective teacher. The recent MET project, funded by the Gates Foundation in the USA, is proof of how important an area of research this is considered to be for educators on an international scale. Besides, better for teachers to be engaged in this discussion and actively contributing rather than leave it in the hands of policy makers to decide. I agree with Joe’s point that one of the most important questions we need to be asking is how observation can be used as a tool for improving teaching and a third of my book is dedicated to exactly that focus. I have argued previously here and here that the single most significant obstacle preventing us from doing so is what I refer to as the ‘assessment straitjacket’ that for decades has constrained the perception and implementation of observation in the English education system. Breaking free from that assessment straitjacket is essentially if we are to fully exploit the benefits of observation as a source of evidence.

There needs to be a ‘thinking outside the box’ when it comes to how observation may be used as a source of evidence in the educational arena. Tinkering with prevailing normalised models of observation is, at best, only likely to have minimal impact and offer short-term solutions to longstanding issues. Although removing the graded element would certainly represent a step in the right direction, for example, it cannot be considered a panacea in itself. In a similar vein, recent calls for the abolition of lesson observation from the inspection process are a classic example of ‘throwing the baby out with the bath water’ and as such represent a knee-jerk reaction to a much more complex problem than the one they claim to solve. Ultimately, what both of these strategies fail to address are the deep-rooted political and epistemological issues surrounding the use of observation as a method of assessment. At the heart of any such discussion is the acceptance that the use of observation is not purely an act of pedagogy but one that is underpinned by issues of hierarchical power and professional trust. Until these issues are acknowledged and discussed by education professionals in an open forum, then any attempts at reforming the way in which the sector makes use of observation are unlikely to progress (O’Leary 2014b: pp. 220-221)

Finally, when discussing the reliability of graded observations in his post, Joe refers to the work of @ProfCoe. Prof Coe is often referred to in other bloggers’ posts about observation. As someone who has been researching and writing about lesson observation, I obviously have to keep up to speed with current research on the topic in the UK and internationally, yet I had never come across any research on observation by Rob Coe. Just to make sure I hadn’t missed something, I tweeted him in March earlier this year and as you can see from his response he openly admits to having done ‘no proper research’. The references to Coe’s work are thus based largely on a powerpoint presentation and a blog rather than empirical research. If people are interested in knowing more about current research in the UK – then this research study, carried out in the Further Education sector, is a good place to start as it is the most extensive research into the topic carried out to date.


O’Leary, M. (2014a) Classroom Observation: A Guide to the Effective Observation of Teaching and Learning. London: Routledge.

O’Leary, M. (2014b) ‘Power, policy and performance: learning lessons about lesson observation from England’s Further Education colleges’. Forum, Vol 56(2), pp. 209-222.

Embracing expansive approaches to the use of lesson observation

Embracing expansive approaches to the use of lesson observation

(This article first appeared in CPD Matters/InTuition – IfL, Issue 8, Summer 2013, pp. 21-22)


In last summer’s issue of CPD Matters I discussed the topic of graded lesson observations in further education and argued that the continued emphasis on measuring teacher competence and performance via the Ofsted 4-point scale had not only become a perfunctory, box-ticking exercise in many colleges, but had also given rise to a range of counterproductive consequences that were impacting negatively on the professional identity and work of tutors in the sector (O’Leary 2012).

In that article I used the juxtapositional terms ‘restrictive’ and ‘expansive’ to describe those approaches to observation that hinder or help professional learning and development. Much of the discussion focused on examples of restrictive approaches and their impact on practitioners, which meant there was less room to discuss the features of expansive approaches. It is to this important area that this follow-up article turns its attention, as I look to present contextualised examples and reflect on why the adoption of a more expansive approach to the use of observation in FE is likely to yield more meaningful and sustained improvements in the quality of teaching and learning than current performative models that continue to dominate the sector.

Defining features

Given the brevity of this article, I have decided to limit my discussion to three specific features:

1) Differentiated observation

2) Prioritising feedback and feed forward

3) Removing the graded element

Space does not allow for a detailed discussion of these three features, but you should at least be able to get an overview. For a deeper exploration please see examples of my other work (e.g. O’Leary 2013; 2014 – listed at the end of this article).

  1. Differentiated observation

Differentiated observation runs counter to conventional models in that it involves identifying a specific focus to the observation rather than carrying out an all-inclusive assessment based on a generic template, as is currently the norm. The observee is given greater ownership and autonomy in deciding the focusand negotiating which session they wish to be observed. The purpose and context thus shape the way in which the focus is decided. So, for example, in the case of the trainee or less experienced teacher, it might make more sense for the observer to play a more decisive role in the focus than they would if they were observing experienced colleagues. What are some of the advantages and reasons for using a differentiated approach to observation?

First, a differentiated approach is built on the premise that each teacher is likely to have differing strengths and weaknesses in their pedagogic skills and knowledge base. Just as the most effective teachers differentiate in their teaching, so too does it make sense to apply this approach to the way in which teachers’ practice is observed. Second, maximising teacher ownership of the observation process is an important feature of facilitating professional learning that is likely to endure. All teachers have a responsibility for their continuing professional development and they are likely to value this more highly if they feel they are given some ownership of the decision making process. Third, the collaborative nature of professional learning means that it is not an individual act or the sole responsibility of the teacher but one that involves colleagues working together. So, for example, there may be times when the focus of differentiated observation is driven by wider objectives across a department such as a departmental improvement plan. These objectives may stem from a range of sources e.g. self-assessment, inspection reports, appraisal meetings etc and may be divided into separate strands or themes (eg use of formative assessment, use of ICT, behaviour management) to address through observation. In this instance a team/department of teachers may choose particular themes to focus on.

  1. Prioritising feedback and feed forward

Feedback is arguably the most important part of the observation process as it is generally regarded as having the most tangible impact on professional development. In a previous research study I carried out, three quarters of respondents across 10 colleges said that feedback lasted no longer than 20 minutes. It is difficult to imagine a professional dialogue of any substantive consequence occurring in such a short space of time. But why is so little time given to feedback if it is recognised as being such an important part of the observation process?

The simple answer is that the time available for feedback and professional dialogue is squeezed because so much time is spent on the collation and completion of the accompanying paper trail and performance management data associated with observations. This is further exacerbated by insufficient time being allocated to the observation process from the outset in many colleges. Feedback, occurring towards the end of the process, invariably ends up losing out. But there are more long term gains to be made from allocating adequate time to feedback in the observation process.

My research has found that those colleges that attach as much significance to the feedback and feed forward stages as they do to the observation itself are often the most successful in improving the quality of teaching and learning, along with fostering a culture of continuous and collaborative improvement amongst their staff. What those colleges have in common is the fact that the importance of feedback and feed forward is not just paid lip service to in their observation policies, but is enacted in practice by allocating appropriate time remission on staff timetables in each academic year.

  1. Removing the graded element

One of the biggest obstacles to embracing an expansive approach revolves around the issue of grading. My research identified correlation between an overreliance on using lesson observation grades as a key performance indicator and low levels of trust and professional autonomy in some colleges. Yet when the graded element was removed, levels of trust between colleagues improved and some of the negative associations surrounding observation vanished, as illustrated in the extract from a research interview with two observers below:

Abdul: We started to not give numerical grades as we felt people concentrated on the number not the feedback and we felt that that worked really well but then the principal decided one day that Ofsted wouldn’t like that and everything came to a halt. We have now moved completely away from that again and everything is performance driven and that’s a shame because that’s where I think we made all of our advances in improving the quality of teaching by getting people on side, being formative as opposed to punitive.

Molly: We did it for just under a year and the impact was quite startling. The quality of learning that was going on rose because staff listened to the developmental feedback rather than focusing on ‘oh I’ve got a three’. We had got staff on side with observations and they were no longer terrorised of having someone in the classroom. They became far more accepting but like Abdul has just said, all that progress has been undone now with the return to grading.

The idea that the summative element can overshadow the formative feedback is well documented in the field of assessment. The grade can take on such importance that it threatens to undermine the value of feedback and the professional dialogue. Abdul and Molly’s account reveals how removing the graded element can be liberating and help to break down some of the negative barriers (i.e. anxiety, fear, suspicion etc.) associated with observation. In their case, it enabled them as observers to gain the trust of tutors and to engage in meaningful, collaborative work, which subsequently led to improvements in the quality of teaching. By concentrating on the feedback and not the grade, the formative aspect of the observation process took on a greater significance and tutors were more disposed to engaging in professional dialogue about their practice.


The way in which staff experience and engage with the use of observation is inevitably influenced by the teaching and learning cultures of the institution itself. The commitment of senior management to promote particular notions of professionalism and professional learning is crucial in establishing an institutional ethos towards observation, which is cascaded, both implicitly and explicitly, to observers and observed alike. The key question for senior managers to consider is therefore a very simple one: what kind of culture do I want to foster amongst staff when it comes to the use of observation? Expansive or restrictive?


O’Leary, M. (2012) ‘Time to turn worthless lesson observation into a powerful tool for improving teaching and learning’. InTuition/CPD Matters – IfL, Issue 9, Summer 2012, pp. 16-18. 

O’Leary, M. (2013) Expansive and restrictive approaches to professionalism in FE colleges: the observation of teaching and learning as a case in point. Research in Post-Compulsory Education, 18(4), pp. 348-364.

O’Leary, M. (2014) Classroom Observation: A guide to the effective observation of teaching and learning. London: Routledge.

Coaching for sustainable development or just working on the observation profile: What are we really doing in the FE sector?


Current challenges

In the FE sector, are we coaching teachers with real development in mind or just to move them from one observation grade box to another on our spreadsheets? This may sound harsh but conversations with coaches in a range of colleges have highlighted this concern and made me feel somewhat troubled at the direction of travel. With increasing pressure within the sector to accelerate improvement, it is easy for coaches to feel that it is imperative they help their coachees to secure that magic grade two, which is taken as a sign of “coaching success”, of the teacher “having improved”. This can lead to an almost exclusive focus on fixing the “faults” seen in the lesson that was graded as a three or four, to the exclusion of deeper, more reflective work on developing the teacher’s practice.

To me, this seems to be a misguided use of coaching…

View original post 1,351 more words

Commentary on #ukfechat forum discussion on graded lesson observations – 28th February 2014

Commentary on #ukfechat forum discussion on graded lesson observations – 28th February 2014

As I lay recuperating in my sick bed last night, I decided to ‘observe’ from afar as a vibrant community of FE Twitter folk debated the #ukfechat topic of the week: ‘Observations: Is it time to ditch the grade?’ As someone who has been actively researching, talking and writing about the topic of observations for the last decade both in the UK and abroad, I was tempted to get involved but my sinusitis persuaded me otherwise that it was best to remain on the peripheries of the discussion as an ‘insider looking in’, if you know what I mean!

As all those working in FE and indeed schools will know, lesson observation is a hotly debated topic. The fact that there was such a lively and diverse debate on last night’s forum should therefore come as no surprise to anyone. In some ways the debate was a microcosm of a wider discussion that continues to reverberate around the corridors of colleges and schools across the country. I recently had a memorable first-hand experience of this when I was analysing and writing up data from the largest study ever to be conducted into lesson observation not just in FE but in the English education system as a whole. In the first part of the project, participants were asked to complete an online survey, at the end of which was an empty box for them to write any comments they had about observation in general. Oh my, did I underestimate the volume of responses that small box alone would generate?! Just under half of all those completed the survey (approx.. 4000 in total) chose to write detailed comments, which when added together totalled over 100,000 words. So, let’s just say there’s no shortage of opinion when it comes to the topic of observations. 

I’m conscious that this is a blog entry and I don’t want it to turn into a long piece of academic writing, if people are interested in that type of thing then they can look at some of the articles I’ve written recently or better still, buy my book! So let’s return now to the forum discussion. I made some notes early this morning of things that stood out for me and I just want to touch on some of those things, not necessarily in any particular order.

The ‘assessing learning/the lesson’ myth

One of the issues that cropped up on several occasions was the old cliché of ‘assessing the learning not the teaching’. For some time now, we have been sold this spurious argument with graded observations that it is the ‘learning’ in the lesson that is being assessed and graded and NOT the teacher. This is a complete fallacy and it needs to be put to bed once and for all. Firstly, if it is the ‘learning’ in the lesson that is being judged, then why does the grade follow the teacher? Why are teachers labelled as ‘outstanding’ or ‘inadequate’ and rewarded or reprimanded accordingly? This is a divisive practice that is commonly reinforced by some employers explicitly naming their ‘outstanding’ teachers, even celebrating their achievements in ‘awards ceremonies’. Besides, if the emphasis is meant to be on the learning taking place rather than the individual performance of the teacher, why are the outcomes of graded lesson observations directly linked to capability procedures in some workplaces?

Any attempts to separate the act of teaching from learning are not only artificial, but crudely ignore the symbiotic relationship between the two. As Ted Wragg (1999) once proclaimed, ‘the act of teaching is inseparable from the whole person and to attack the one is to demolish the other’ (p. 91). And the idea that ‘learning’ can be accurately measured through the medium of observation is highly contested and the reality is that we are light years away from ever being able to make such a claim with any degree of authority. The pseudo-scientific art of grading seduces us into believing that observer judgements have greater objectivity and reliability than they can actually claim to have. And why is that? It is because on the surface numbers have a ‘scientific’ quality to them, which makes people less likely to question what they are deemed to represent. In the case of graded observations, there is an assumption that the use of the Ofsted 4-point scale has some kind of objective value comparable to the use of a calibrated measuring instrument such as a thermometer. Yet this is clearly a myth. They are, of course, dependent on the subjective interpretation of observers so the application of a grade can never be wholly reliable. 

Wanting to be graded or know the grade

Of course, there are some teachers who are keen to want to know the grade even if they’re not being graded. This is indicative of what I’ve referred to in previous work as ‘normalised behaviour’. In other words, such teachers have become institutionalised into expecting a grade to be attached to an observation, regardless of the context or approach. They are unable and/or unwilling to conceptualise the use of observation outside of a performative context and see an umbilical link between their classroom ‘performance’ and attempts (because that’s all they are) to measure it. I can understand the ‘reward’ incentive of this for some but I think such a mentality does little to foster a collegial and collaborative culture in the workplace. I’m not opposed to the notion of competition per se but firmly believe that there is a time and a place for it and this is not it. 

Tweeters that stood out

Overall I thought the level of debate was fantastic and I enjoyed observing it from afar. I feel there are a few tweeters who need a brief mention though as for me their balanced and critically reflective positions shone through their tweets and they were: @hannahtyreman @Shanie_Nash and @cazzwebbo. Also @GrahamRazey deserves a mention for raising the all-important point about cultures of teaching and learning being the lynchpin of any successful model of observation.


Concluding thoughts

What we need is a fundamental reform of the way in which observation is used. Tinkering with the present system is pointless and only likely to have a minor impact. At the heart of such fundamental reform is the need to reconfigure the contexts and cultures of teaching and learning in which observation occurs, as Graham so rightly alluded to in his tweets. It’s not just about moving from one formulaic model to another but root and branch reform; a fundamental reconceptualisation of how we engage with observation and that inevitably requires removing the assessment straitjacket that currently constrains how people perceive it and what it’s used for.