Read my ideas about education, politics, language and society. I have included some autobiography, and considerations of what it is to be a man in his seventies in rural England.
Tuesday, 29 December 2009
Have yourself a merry/melancholy/some other emotion Christmas
This mixture of emotions explains the persistence of Christmas songs that express regret and disillusion, like The Pogues’ Fairytale of New York, and Greg Lake’s I Believe in Father Christmas. Another popular song, Have Yourself a Merry Little Christmas from the film Meet Me in St Louis, differs in being melancholy because the singer looks ahead pessimistically to future Christmases. This made the lyrics unacceptable for those who dislike emotional complexity, and new, more upbeat, ones were devised for Frank Sinatra to sing. This re-writing follows another pattern of Christmas - the hope that over-eating, alcohol and jolly songs will swamp any emotions we find painful. Spending becomes a way in which we can avoid understanding ourselves.
Thursday, 17 December 2009
The Privatisation-IT-Consultancy Complex
New technology is promoted with particular enthusiasm by the military-industrial complex, ostensibly because it offers the possibility of gaining an advantage over prospective enemies. But new technology is usually less reliable and more expensive than the equipment it replaces. The more unreliable the technology, the longer the contracts, which offers the prospect of greater profits since contracts for military equipment are awarded on a cost plus basis. And equipment failure and cost over-runs can always be explained to the public as an inevitable risk required for the effective defence of the state.
In British central government, domestic expenditure is far more important than spending on the military. Our own equivalent of the military-industrial complex therefore is built around on the vast sums spent on health and social care, public order, and transport. Public funding for these areas can be seen as a treacle pot from which members of the complex can feed, and there are three ways in which they can do so.
The first is through the privatisation of what were previously regarded as being government services. This has expanded vastly since a neo-Thatcherite government came to power in 1997, and has been rationalised as introducing competitiveness and the efficiencies of the private sector into public services. A key component of privatisation is the Public Finance Initiative (PFI), which involves public authorities paying contractors high fees to rent from them newly-built hospitals, schools and other public buildings (usually with additional payments for servicing the buildings). Besides PFI, there are contracts for transport, prisons, nursing and residential care, cleaning services in public buildings, employment services and so on. Needless to say, finance raised to pay for PFI and related schemes has involved large transfers of public funds to the banking sector.
The second way is through consultancy fees. These have expanded with the introduction of contractual relationships throughout government, usually following privatisation. One or more of the big four professional services firms (PWC, KPMG, Ernst & Young, and Deloitte Touche Tohmatsu) appear in every such transaction, and are even brought in by government to audit each others’ work, and report on each others’ failures. They have managed to sell a spurious technology of management expertise to politicians and the public, such that no major governmental decisions can be made without their engagement.
The third way is through major IT projects. These have involved the world’s most expensive database (the NHS Programme for IT) which will probably have a final cost of £20 billion, plus such projects as the database for identity cards (£12 to 18 billion), ContactPoint (the database recording all children, costing £1 billion), The Libra court management system (£341 million), and the C-NOMIS offender management system (£279 million). All of these systems have spectacular cost overruns, and have failed to deliver the benefits they initially promised. Very large parts of the NHS Programme for IT are hardly even used.
The complex can therefore be referred to as the ‘Privatisation-IT-Consultancy Complex’, or ‘PIC Complex’ for short. As with the military-industrial complex in the USA, the PIC Complex in Britain shapes government policy in its own favour. There is no public demand for identity cards, and security experts see them as having a negligible impact on public safety. The phenomenal cost of the scheme is seen as an argument against it. But the very cost of the scheme is the reason it exists - it will produce an endless stream of public funds from the treacle pot for IT companies, professional services firms, and the financial sector. Few if any experts in child care asked for a child database, but we have paid for one which will almost certainly not work, and will divert resources from services which do protect children. The diversion of resources from services which deal with people into IT schemes which enrich the PIC Complex has already happened with C-NOMIS.
How can this be allowed to happen? One reason is that the concentration of power in English government gives great leverage to the PIC Complex. Its members only need to convince a small number of senior politicians and civil servants to secure multi-million pound contracts. This saves all the bother of selling products to lots of local councils, hospitals schools etc, let alone having to convince the public who ultimately pay for them. This is a symbiotic relationship - privatisation and large centralised databases weaken rival local centres of power and hence maintain this leverage. Convincing small numbers of senior politicians and civil servants has proved easy because some are for sale. This is not traditional corruption in which a contractor pays money and then gets what they want, but a new time-displaced version in which the politician favours a contractor and is subsequently rewarded with a seat on the board, consultancies etc. I would not of course allege such a thing about Patricia Hewitt MP, the former Secretary of State for Health, who is currently earning £12,500/month from BT Group (a major contractor for the NHS Programme for IT), and £4,600/month from Alliance Boots (which is a major pharmaceutical distributor to the NHS and a contractor for privatised primary care services). October was a good month, when she also earned £15,000 from Cinven, which owns a chain of private hospitals which contract with the NHS.
Academics are also subsidiary players in the PIC Complex, providing essential legitimation services, using suitably high-minded and abstract terms like ‘contestability’ and ‘choice’. But more of that in a later posting.
Thursday, 3 December 2009
Learning our incompetencies
Learning our incompetences helps us avoid wasting time on activities where there are more talented rivals, and enables us to concentrate on what we can do effectively. However, there are some areas of life in which this learning usually fails to take place. Most people, against all the evidence, continue to believe they are ‘above average’ drivers and lovers. Many people also believe that, despite lack of training or any previous evidence of competence, that they can run a pub or a restaurant.
The results of this delusion can be found in every town. Here is an example. My wife is a member of a group of women (which she calls the ‘ladies’) who have worked together in the past, and celebrate each others’ birthdays by meeting for dinner. The restaurant is usually one chosen by the lady with the birthday. The most recent meal was at a new Indian restaurant in Worcestershire. The ladies met there at 8pm. The decor was pleasant, and there were poppadums and chutneys awaiting them at the table. But no waiter came to ask for their orders for drinks, and there was a long delay before any waiter came at all. When this did eventually happen, the ladies decided to skip a starter and ordered main courses. Nothing happened for a very long time. The group at the next table decided that they too had waited for far too long, and decided to leave. The waiter pursued them into the street and told them the food was ready. The food then came for the ladies. It was lukewarm, indicating that it had been cooked some time ago and then forgotten.
No waiter came to ask them if they wanted to order a dessert, so the ladies decided that they would move directly to coffee. A coffee pot arrived, but no cups. The ladies found some in the restaurant and served themselves. They then seized a waiter, who produced three bills. These included one for £27 worth of lager, which they certainly had not drunk. There followed a process of negotiation to reduce the bills. The toilets were in the basement, but when reaching them, the ladies found they were closed, and they were redirected to the disabled toilet at the top of the building (!) By the time they were due to leave it was 11-30pm, and the front door of the restaurant was locked.
The sad thing is that the ladies could have gone to the Spice Cuisine in St John’s in Worcester and paid rather less for an excellent meal, cooked rapidly and served efficiently. But then, the people who run and work in the Spice Cuisine have discovered an area of life in which they are supremely competent. The staff at the other restaurant could try being footballers.
Wednesday, 25 November 2009
The Hadamar Clinic and humanity
The policy of mass murder of people with an intellectual disability by the Nazi regime was an extension of a widely-approved public policy of eugenics. In the first part of the 20th Century, eugenics attracted support from left and right, and from leading intellectuals. Indeed, it was the more socially-progressive societies like Sweden and Canada which at that time most ardently promoted the eugenic policies of the compulsory sterilisation of people with an intellectual disability, and their incarceration in mental handicap hospitals. Eugenics became discredited by its association with the Nazis, and public policy in almost all countries today favours the integration of people with an intellectual disability into the day-to-day life of society. The old hospitals have closed, the rights of disabled people are increasingly protected, and strenuous efforts are made to improve their employment opportunities.
It is tempting to see these changes as marks of ‘progress’, in the sense of an incremental improvement of civilised values from barbarism to humanitarianism. But this would be historically inaccurate. The harsh policies associated with the eugenics movement replaced those influenced by more humane ideas in the 19th Century, which had emphasised the potential of all people with an intellectual disability for learning and social improvement. Rather than a march of progress, our recent history has been a struggle between two views of humans and their worth. Eugenics represented the idea, popularised in the Enlightenment, that human beings are distinguished from other species by their rationality. Rationality then becomes the measure by which people can be ranked, but also a means of determining the general arrangements of society. This legitimises ‘social engineering’, or the application by those in power of measures to shape the lives of those deemed less rational than themselves. A fear in the early 20th Century that the less rational sections of the population were increasing in numbers compared to the more rational created support for eugenic ideas, and led to the ultimate Nazi social engineering project at both the Hadamar Clinic and the mass industrial-scale killings that followed.
There is another and older view of humankind: that each of us possesses an essential essence or soul, that gives us our human character and by which we can be judged. The quality of our souls is unrelated to our physical strength, our intelligence, or our rank in society: a person with an intellectual disability can have a soul more worthy than that of a scientist that sneers at him. This is of course a fundamentally religious outlook, which should make us very troubled by attempts to drive out religion in the name of science. What, therefore, should be the duty of the scientist and the intellectual according to this older view of mankind? It should be to find out the truth and tell it to others, to maintain knowledge in society from one generation to another, to help people reflect deeply on their values and choices, and to do all this with humility.
Tuesday, 17 November 2009
The curse of the generic
The decision to merge these departments followed the ‘Seebohm Report’, which correctly noted that some families were involved with two or three of these agencies, and incorrectly proposed that it would be more efficient to have a single generic ‘family’ service. The resulting merged social services departments were large and had management hierarchies rather than being led by a senior professional. The commitment to ‘generic’ social work, in which each member of staff dealt with the full range of clients, became departmental orthodoxy. Both these trends led to a rapid exit of the most skilled and senior staff. They were replaced during my first year as a social worker by people like me: well-meaning, untrained and incompetent.
The results across the country were a radical decline in the quality of child protection, and support for disabled and mentally-ill people. The first indicator of this was the avoidable death of the child Maria Caldwell. The subsequent official enquiry identified that a major cause of institutional failure was the confusion among social workers about whether their primary responsibility was to the child or to the ‘family’ (ie her parents). This was the first of many such enquiries, which led to a succession of management ‘solutions’, from inter-agency committees, registers of children at risk, centrally-imposed targets, inspections, child databases, and repeated re-organisations. No-one in power paid much attention to enhancing the professional skills of social workers involved with children, enabling them to develop specialist skills, or setting up the kind of small specialised and professionally-run departments that had been a success in the past. When specialism did arrive, it was implemented as part of a bizarre governmental reform which merged local authority child protection services with local education departments.
Why this resistance to specialism? I think it is a product of the managerial control that arises with the creation of large public organisations. In small organisations, staff are known as individuals, and there is an awareness of their different strengths. Staff can be assigned to different work informally, and their supervisors can generally assess their performance by observation and informal meetings. Large organisations see staff as a block of people to be matched with some quantitative indicator of workload. It is easier to move people around if they are supposed to have generic responsibilities rather than diverse specialist skills.
The drive to generic work and consequent de-professionalisation arises in many large public organisations. This can occur even in organisations in which specialist professional skills are regarded by almost everyone as being essential. The National Health Service has attempted to grade all its diverse professions on a single ‘knowledge and skills framework’ - a spectacular example of the kind of ‘blue skies’ (crackpot) thinking that occurs in very large organisations. At the same time, the government has attempted to reduce the time spent in specialist medical training. There are similar trends in universities. These value their most highly skilled staff, at least as long as they attract large research grants, but post-doctoral researchers and academic staff who specialise in teaching are sometimes treated as classes of helots, interchangeable and disposable.
The curse of the generic partly explains an apparent paradox: as public organisations get larger and employ more managers, the less competent they are in delivering effective public services. There are other explanations for this paradox: the conversion of previously-autonomous professionals into highly-regulated functionaries produces the alienation familiar in industrial process work. Also, long management hierarchies move decision-making further from the organisations’ customers, who usually encounter junior members of staff with limited authority to adapt procedures to meet individual needs.
We therefore need revolutionary change - towards small-scale public services, with a re-assertion of professional specialism and autonomy. We need to down-size schools with thousands of pupils, so-called ‘local’ authorities which cover wide areas of the country and multiple and formerly self-governing towns, and large welfare departments which fail to protect children at risk or adequately support the disabled. Of course, some public agencies will always need to be large: big cities need governments, and the large numbers of students in higher education will probably require large universities. However, authority can be devolved within cities to community councils (as in Scotland and Wales), while universities can operate more on the Oxbridge model, with academic staff working in semi-autonomous colleges. After all, Oxford and Cambridge Universities have hardly been failures despite lacking the benefits of centralised management.
Thursday, 5 November 2009
Disneyfying the body
This tendency seems to be an extension of our innate ability to empathise with each other - an ability said to be lacking among people with Autism. It is one step from understanding animals as if they are human to telling stories of them as humans, speaking and wearing clothes. I call this process ‘Disneyfication’ after Walt Disney, who set up theme parks full of humans pretending to be animals, to resemble cartoon characters of animals resembling humans.
Disneyfication does not stop with animals. Genes can be Disneyfied too: Richard Dawkins has sold a lot of books called ‘The Selfish Gene’. I went to a presentation this week on techniques for regenerating cells in the Central Nervous System. The researcher spoke unselfconsciously about neural cells ‘choosing’ between options, and ‘preferring’ one binding site to another. I am sure a Disney or Pixar cartoon of neural cells, dressed as people, falling in love, and plotting with each other to rebuild a brain will arrive soon at local cinemas.
Tuesday, 3 November 2009
Surviving school
The ethos of the school was incomprehensible because it involved a strange nostalgia for an imagined mediaeval England. The headmaster renamed his post ‘chief master’ to revive some supposed tradition, and the school was organised into ‘houses’ to mimic private schools. The school’s sports teams only ever played private schools such as Bromsgrove School, Malvern College and the like, and never the oiks of such schools as King Edward’s Camp Hill. Around the school were boards with the names of old boys who had gone to various Oxbridge colleges. London University was given a board on its own, and there was one for ‘other universities’. Classics was the most esteemed academic subject, and there was a sneering approach to anything practical. I completed a science project on computers (then in their very early days), and the teacher dismissed it for being too technological. There were endless petty distinctions and grades among the students, all to create a world of dull conforming hierarchy. So a school in a metal-bashing city famous for its inventiveness and outspokenness produced students equipped to thrive in a synthetic medieval nostalgia.
I left the school with relief at the age of 18 and went to Handsworth Technical College and completed three social science A levels in a year. My life opened out: I enjoyed the chance to learn as an adult, study things I enjoyed, and meet people from different cultures. I then went to the London School of Economics and studied at a time when English universities were at their peak, full of new ideas about education and the world, taking students from a wider social background than before, but not yet swamped by the vast numbers who come to university now. In those days, undergraduates were tutored by professors and the leading academics of the day: mine included Alexander Irvine (later Lord Chancellor); Edward Mishan (the first economist to challenge the worship of growth); and an American lawyer William Letwin (father of the Conservative politician).
Now my life has come full circle. I work in a university about a quarter of a mile from King Edward’s School. It takes me a long commute of an hour to reach work from home. The University is thankfully more aware than King Edward’s School of the need to be in the forefront of knowledge and to engage with its City. But I miss the intellectual challenge I encountered at the London School of Economics, and the hope for a better future which inspired me when I was a younger man. I was unhappy with the synthetic mediaeval nostalgia of King Edward's School, but I can now see that, for its teachers, it was an attempt to maintain the continuity of the human spirit after the terrible years of war they had experienced. Their beliefs, however old-fashioned, were in any case superior to morality of money and power that eventually triumphed in England.
Monday, 26 October 2009
The rule of anarchy
Last came Anarchy: he rode
On a white horse, splashed with blood;
He was pale even to the lips,
Like Death in the Apocalypse.
And he wore a kingly crown;
And in his grasp a sceptre shone;
On his brow this mark I saw -
'I AM GOD, AND KING, AND LAW!'
With a pace stately and fast,
Over English land he passed,
Trampling to a mire of blood
The adoring multitude
Shelley’s association of anarchy with the supreme authority and power of the state differs from most popular uses of the word, in which ‘anarchy’ is used to designate a disorganised powerless multitude in which private appetites are fulfilled by force. Shelley’s view (and mine) is that most people, left to themselves, will organise their lives peacefully. The main causes of violence and hatred are the abuse of power and the love of power respectively. These are more likely to emerge from those who have already accumulated power, seek more, and fear its loss.
The control of the powerful has therefore become the major political and legal enterprise in any society. This takes the form of ensuring decision-making is shared with representatives of those affected by the decisions, that proper deliberative procedures are followed, and that decisions are implemented on the basis of clear rules rather than arbitrary favours to cronies and courtiers. Needless to say, societies which operate in this way tend to make better decisions which command greater consent. They also tend to be more prosperous: people will work to accumulate money if they can be assured it will not be seized from them by some arbitrary decision.
These principles are well-understood when people think of societies and governments, but not when they think of big organisations. Corporations, banks and universities can all experience the rule of anarchy. Rules may exist in the organisation, but they are by-passed by the powerful. Off-the-cuff decisions replace due process and the thorough discussion of options. Arbitrary favours are handed out to cronies and lovers. The central aim of the organisation becomes the accumulation of prestige, wealth and power for the small ruling junta. Such organisations, like some banks recently and many commercial organisations in the past, make increasingly grandiose and inept decisions, and crash to disaster. Commentators are then amazed that such apparently powerful and dominating figures - like Fred Goodwin (formerly of the Royal Bank of Scotland) or Kenneth Lay (formerly of Enron) - should fall so suddenly. To appreciate their achievements, we need to return to Shelley:
I met a traveller from an antique land
Who said: "Two vast and trunkless legs of stone
Stand in the desert. Near them on the sand,
Half sunk, a shattered visage lies, whose frown
And wrinkled lip and sneer of cold command
Tell that its sculptor well those passions read
Which yet survive, stamped on these lifeless things,
The hand that mocked them and the heart that fed.
And on the pedestal these words appear:
`My name is Ozymandias, King of Kings:
Look on my works, ye mighty, and despair!'
Nothing beside remains. Round the decay
Of that colossal wreck, boundless and bare,
The lone and level sands stretch far away"
Tuesday, 20 October 2009
Great crackpot ideas of the past
The ‘nation-state’ is a great crackpot idea because the peoples of Europe and the rest of the World have not distributed themselves into neat geographical clusters. Of course, rulers have tried for centuries to make their subjects more homogeneous, usually by attempting to eliminate inconvenient groups that persisted in adhering to minority religions, languages or customs. The ‘nation-states’ which dominated Europe after 1918 sought even more vigorously to make reality fit the crackpot idea, with mass transfers of population, suppression of minority languages, and exterminations. The creation of the European Union is an admission by all but a few recalcitrant nationalists in Europe that the nation-state has been a catastrophe for the continent, and that a muddled unity is preferable.
There are lesser crackpot ideas too. These cause havoc and dismay but do not usually involve major loss of life. An example is the research assessment exercise (RAE) for allocating funds in British universities. This allocates about £2 billion/year, according to a grading of ‘research excellence’ across all disciplines. These grades are assigned by a series of expert panels based on their assessments of published research papers submitted by universities, and on various intangible fudge factors like ‘research impact’. RAE assessments are completed every few years. The total cost of this exercise has been estimated at £47 million, mainly in the time spent by university staff in preparing submissions for the expert panels.
What makes this a crackpot idea is not just the high transaction cost of the exercise, but the impossibility of grading all human knowledge from genetics to philosophy on a simple rating scale. Most academics could probably agree about the leading centres in their discipline, but there would be probably be little reliability in any ratings beyond that. To complicate matters, there is some excellent research in some mediocre university departments and vice-versa. Academics would also struggle to reliably rate research which crosses traditional disciplinary boundaries or which challenges dominant paradigms. Watson and Crick were lucky to have done their research before the RAE existed.
To make matters worse, the RAE grades past research, not current performance. Because of the long intervals between assessments and the delays in academic publishing, some of the research assessed may be based on laboratory or fieldwork almost a decade old. Panel ratings are therefore exercises in the history of knowledge. There is an interesting contrast when academics bid for funds to carry out new research projects. In this case, the research councils and the various charitable research funding agencies assess the academics’ recent performance and the prospect of a useful outcome from their proposed research project. Unfortunately, research funds are limited, and academic staff spend much of their time preparing research bids with limited chance of success.
The obvious solution to all this would be to close down the RAE, and transfer the funds to the research councils. This would also release the high transaction costs of the RAE. Why won’t this happen? The answer is that however crackpot an idea, it generates its winners and losers. The universities which gain most funds from the RAE worry they might lose by an alternative system. Besides funds, the RAE assigns esteem, and esteem counts for a lot in the competitive and hierarchical world of universities. Central government is also a winner: it can enforce regular changes in the formulas used to calculate the allocations of funds following the RAE, and thereby remind those working in higher education of their subordination to the state. There are important lessons of power here: use complicated rules to disguise the exercise of your power and to divide those you rule into squabbling factions. Get clever people to spend all their time in silly competitions, and they will pay less attention to the serious business of challenging your power.
Thursday, 8 October 2009
Bullying as a career
Some years ago, the Government tried to improve the recruitment of teachers with the slogan ‘You never forget a good teacher’. That may be true, but you never forget really bad ones either. Most of all, you never forget the school bully. Bullying at school causes absenteeism, illness, and even suicide. Even after leaving school, the pain can live on. I know of friends who have had chance encounters with school bullies years later, and felt the same daggers of pain, anxiety and humiliation.
School bullies are usually portrayed in fiction as thick and cowardly, who inflict cruelty because of psychological abnormalities. In fiction, they are defeated in the end. But this is wishful thinking. An alternative fictional school bully is shown in Michael Palin’s Tomkinson’s Schooldays. Here, the school bully enjoys the exercise of cruelty, but uses it to gain exceptional favours including access to cigars, whisky and the pleasure of attractive young Philippine women. In Tomkinson’s Schooldays, being a school bully is an important career, leading directly to the Cabinet.
Bullying has been a profitable career for many others. Those who enjoy personal power over others and exercise it cruelly will not only succeed in life, but will usually accumulate an adoring circle of cronies. This is because bullying serves many functions. Apart from the obvious gains of encouraging compliance, it can generate solidarity. In mediaeval Japan, the shoguns created a caste called ‘burakumin’, who were assigned the most inferior status in society. All others could share the joyful common task of bullying and humiliating them. The British Conservative Party has a successful history of building support by identifying groups of victims who can not fight back, from immigrants, to single mothers, to the chronic sick on benefits.
In some cases, bullying is a response to specific impediments to management. An example would be in universities, where many staff are on ‘open contracts’ and can not be dismissed except for gross misbehaviour. University leaders wish to enhance the prestige of their institution (and hence themselves) by expanding research. Research is deemed to be an exceptional intellectual skill, not available to people who are committed to lesser forms of scholarship such as teaching. University leaders therefore aspire to recruit researchers in place of teachers, and offload teaching to even lowlier staff. But if teachers are open contract with a full workload, they can not be made redundant. Bullying is therefore employed. This can take the form of denigrating their work and closing the courses they run, excluding them from senior positions, allocating them to inferior work spaces, and threatening disciplinary action for minor (or no) infractions.
Heaven forbid that you might think this true of my own dear university, led as it is by saintly figures thinking only of the welfare of the staff and students in their charge.
Tuesday, 29 September 2009
The generation of heroes and the generation of would-be heroes
Seeing The Ipcress File also made me reflect on why I wanted to be a hero in the first place. Born in 1946, I was raised in the shadow of war. My parents and all the members of their generation had stories about the War. One of my uncles had been in the Eighth Army in Africa and Italy, and another in the Airborne forces from D-Day onwards. My father had been unable to serve on health grounds, but all civilians were on the front line when the air raids began. There were gaps in local streets where bombs had fallen, the centre of town (Birmingham) had ‘bomb sites’. A local wood was full of rubbish left behind by American troops who had been stationed there. Comics, films and newspapers were dominated by stories of the War. Heroes were everywhere. ‘Heroism’ was more than bravery in the face of the enemy: it meant endurance and stoicism, a sense of common purpose, and a willingness to sacrifice for the sake of others. Many members of my generation absorbed these values, even though there was no war left to fight. The substitute for some of us was politics: an heroic struggle against oppression and injustice which involved marching but no gunfire. We followed the campaigns in the USA against segregation as if it was our own battle. We marched against nuclear weapons, and when we came of age, we marched to support Dubcek and oppose intervention in Vietnam.
Of course, our generation of would-be heroes was pushed aside by younger people not raised in the shadow of the War and who responded to the peacetime values of self-enhancement, both financial and psychological. These new values became important in the later 1960s. Sexual hedonism and drugs were followed by searches for mystical enlightenment. Eventually, most people realised that financial and psychological enhancement could best be attained by an endless sequence of purchases. These values came of age politically in this country in 1979, and still dominate the political class. As a result, politicians still talk like members of rival firms seeking to persuade reluctant consumers to buy their products. When faced with real crises, like recession or global warming, they lack the skills or the understanding to mobilise mass support. There is once again a need for politics to be an heroic enterprise, but we have no heroes any more.
Wednesday, 26 August 2009
The Chaos in our High Streets
▸ Currys. Despite my persistent demands at our local branch for a chicken tikka marsala, they insist that they only supply electrical goods.
▸ Boots. A series of mysterious stores which sell make-up, medicines, domestic goods and lots of other things, but no footwear.
▸ Thomas Cooks. The name suggests that these are either restaurants (staffed by people called Thomas), or shops selling kitchenware. All they seem to sell are holidays and foreign currency.
▸ Office. A recent arrival in my local high street in Worcester. Demands for stationery were not welcome, and all they stocked were women’s shoes.
▸ Bank. The most confusing of all. No savings facilities, loans or credit cards, just piles of clothes on sale.
This trend to name shops after things they do not sell must come to an end. Our government must act in the name of health and safety, security, or any of the other reasons they usually summon up to order people around.
Monday, 10 August 2009
The Laws of Information No. 3
3. Data that is collected to measure performance loses validity.
First a confession. In the late 1980s, I worked as Director of Planning and Information in a mental health service in the NHS. One of my tasks was to organise the statistical returns on clinical activity for despatch to the Department of Health. At that time, these were based on a set of standard definitions called the ‘Korner system’ (after a woman who chaired a committee which recommended them). Our service included a brilliant and very hard-working consultant psychiatrist for the elderly. She believed that assessments of new patients should initially be in the patient’s own home (a ‘domiciliary visit’ or ‘DV’). Since she was an orthodox Jew, this meant a lot of walking when her duty days coincided with the Sabbath. Unfortunately, the Korner system required information about scheduled outpatient clinics but not domiciliary visits. Following the Korner rules would have meant that our most active consultant would appear as our least active. This was obviously unjust, so I modified the returns for her clinical activity to record each DV as an attendance at a (non-existent) outpatient clinic.
Paradoxically, my data-adjustment produced statistical returns which were a more accurate reflection of clinical activity than would have been the case without such adjustment. Nonetheless, they became an inaccurate record of inpatient clinics in the service in which I worked. I suspect that data-adjustment in the desired direction was and is rife in the NHS. Although this is dishonest, it can cause far less damage than changing reality to generate honest statistics. A well-known example of changing reality in the NHS is to make patients wait in ambulances outside A&E departments. This reduces the time the patient spends in A&E for the purposes of official statistics, and hence enables the hospital to meet a government target. There are many, many more examples in the NHS of how meeting centrally-imposed targets can damage patient care.
This is not a recent phenomenon. The whole technology of corporate strategic planning and management by targets owes its origins to Gosplan, the state planning commission in the USSR. Studies of the Soviet economy in the 1960s and onwards were full of examples of how rational responses by individual enterprises to centrally-determined targets could produce absurd results. These included the shoe factory that met its target number of shoes by producing shoes all in one size, and the steel factory that met it target for weight of steel by producing a few huge ingots.
Friday, 31 July 2009
The rise of the celebrity wedding
This kind of wedding is increasingly popular, and is modelled on the celebrity example. As social relationships within communities weaken because of commuting (for both men and women) and the domination of television, people take their guidance on how to live their lives from the examples of celebrities. Tradition, religion and morality become less important than an outward display of mimicry. The greatest celebrity events are weddings, held in private so that exclusive rights for photography can be sold to Hello or OK magazines. The event is therefore staged with the care that would be expected of a film or a play in the West End.
The celebrity example means that many people now regard a wedding as an opportunity to star in their own theatrical event. This has inflated the cost of weddings, which have become the largest single item of expenditure for many couples apart from buying their house. People can not usually afford such an event when they first live together, and so weddings are postponed for many years. This has a profound effect on the meaning of the event. A wedding of this kind is no longer a commitment by a man and a woman to live together and support each other through life, but is instead a party to celebrate several years of sustained cohabitation. It is no longer a union of two families celebrated in their presence, or for that matter in the presence of the public.
So church bells will ring less often across the fields of my village on a Saturday, and weddings will become private fancy dress parties for ageing couples.
Monday, 27 July 2009
The Laws of Information No. 2
The cost of information is felt most acutely by staff when it is required from them. There are routine statistics to be completed, forms to be filled on staff and student progress (including one for every single meeting with a research student!), surveys of staff satisfaction, and one-off requests for information which have descended the management line (usually with shorter and shorter deadlines at each stage of transmission). Staff usually see these requests as a chore to be completed quickly, and do not therefore strive tirelessly for accuracy in collecting and recording the required data. This leads to the second law of information:
2. Data is always less reliable than you think.
Scientific texts emphasise the potential pitfalls in gathering data, and careful scientists have standard routines for checking its validity and reliability. Gathering research data from people is particularly troublesome because of their capacity to fabricate, to rationalise, to forget, and even to avoid telling the truth as an act of politeness. Even in a world where people did none of these things, there would still be lag between events occurring and data being collected, inconsistent application of rules for categorising data, and missing data. Yet these limitations are usually ignored when organisations collect and process information from their staff or from the public. Instead of using wide confidence intervals when reporting the information they have collected, organisations glibly report data to an exact percentage point. There are earnest debates about small changes in statistics from one reporting period to another, even though these are probably within confidence intervals.
Interpreting data would be difficult enough if it was simply a matter of general unreliability, but there is the far bigger problem of biassed unreliability. This is the third law of information:
3. Data that is collected to measure performance loses validity.
I will deal with this in the next posting.
Tuesday, 14 July 2009
The Laws of Information No. 1
1. Information is costly. Back in 1968, there were no photocopiers, office computers, or emails. If you wanted a copy of a letter, the typist had to insert carbons and additional sheets of paper when she typed. There was a limit of about three or four copies that could be made this way. If you needed more, then a different process was required. The typist would type the report on specially-waxed ‘skins’, which were attached to the drum of a machine we called a ‘Gestetner’. The drum would contain thick black ink, which you always got on your hands. Both methods of copying were costly and time-consuming, and a major O&M task was therefore to reduce the amount of unnecessary information circulating round the factory. We did this by creating a flow diagram for all the routine reports generated by staff, and asking their recipients whether they found them useful. We found that many reports had begun as one-off requests by management to meet a specific need, but had then become routinised. Some reports went straight from the envelope to the waste paper bin.
This seems a lost world now because photocopiers, word-processing and emailing have successively made the production of multiple copies much easier. But this has had the effect of shifting the cost of information to the reader. People in offices now spend hours a week sifting through emails, most of which come from their seniors but are irrelevant to their work. Emails accumulate in inboxes, and the ones which require rapid attention are missed. Because the idea has taken root that information is cheap to reproduce, staff are required, often at short notice, to produce data and statistics for senior management. As in the past, these requests can become routinised even when the original need for the information has passed.
This indicates that organisations should revert to the O&M principle of reducing the flow of unnecessary information, to release staff time and speed up their response to the information that really matters. Without this, problems develop with the data we do have, which I will look at in a later posting.
Thursday, 9 July 2009
Life as a palimpsest
Before printing on paper was invented, texts were written on parchment made from animal hides. Parchment was durable but expensive, and so it wasn’t wasted. If you had something to write, you took an existing parchment with writing on it, rubbed out a line of old text, and wrote in the space you had made. After this had happened several times, the piece of parchment contained multiple erasures and bits of text which, if read in sequence, made no sense at all. This type of parchment is called a 'palimpsest'.
The word ‘palimpsest’ has been used metaphorically to describe cities. Bits are knocked down and replaced by many different architects and builders, all with different aims in mind. This is particularly true of the sort of European cities in which the streets are not laid out in grids and where no king or emperor has been able to impose an overall plan. Palimpsest cities may make no sense (particularly to a visitor), but can be pleasant to discover: alleys and streets wind in mysterious directions; streets suddenly open out into hidden squares; churches and other imposing buildings occupy sites next to houses and office blocks.
‘Palimpsest’ can also be used to describe organisations. An example is the National Health Service in England, which has had numerous organisational changes and endless new initiatives, each with a new set of organisations to implement them. The resulting organisational structure makes the sort of sense familiar to readers of palimpsests. But the NHS keeps on functioning because the real work is done by doctors, nurses, paramedics and other people who know what they are doing. Chaos only intrudes when politicians, the Department of Health or some part of the senior management interfere. Living in an organisational palimpsest, they naturally speak a higher form of management gibberish (‘targeting the deliverables’ etc). In fact, the decay of language into this kind of gibberish is probably an indication that those who speak it live in a world of meaningless procedures and incomprehensible systems for evading responsibility.
Human life itself could be seen as a palimpsest. As you get older, your memory gets over-written by random experiences, different skills and knowledge. You make off-the-cuff decisions which have major implications for the rest of your life, and make sudden and unexpected changes to what you had intended to be an orderly and planned life. Of course, you don’t see it that way when you look back. Human beings have a marvellous ability to rationalise their actions and to see stories (and even conspiracies) where there are only random events.
Wednesday, 1 July 2009
New gods for old
Living gods are created by their worshippers, but do not usually require any other forms of special behaviour. Their worshippers are not required to forego particular types of food or wear a special set of clothes or be more moral than the rest of us. The latter is particularly convenient because the people chosen as gods rarely lead exemplary lives. However, a true worshipper will deny any reports of their god’s wrongdoing (even pederasty and drug abuse), or regard it as a sign of the trials inflicted on the god to attain his or her status. Worship therefore involves idealising as well as idolising, and the worshipper passionately follows the lives of the god and creates special shrines in their homes of various icons of him or her. When the god dies, death is denied, symbolically and sometimes literally. The god dies in both senses when their last worshipper has passed away. Many years ago, I have met an old woman who still worshipped Rudolph Valentino and another who worshipped Mario Lanza. None surely are left alive who worship these mortal gods now, and Michael Jackson’s worshippers too will pass away over the next decades. People might still go on singing his songs though.
Friday, 26 June 2009
Valuing intellectually abnormal people
This is all justified as being ‘inclusive’ - the latest term in a sequence which began with ‘normalisation’, via ‘social role valorisation’ and ‘ordinary living’. However, these principles to date have only been imposed on people at one end of the IQ scale - those who are two standard deviations or more below the mean, equivalent to an IQ score of lower than 70. To avoid discrimination, we should of course apply the same thinking to the other group of intellectually abnormal people: those two standard deviations above the mean, with an IQ of 130 and above.
When we look at the life-styles of this latter group in England today, we can see that these too offend the principle of inclusion. This is particularly true of arrangements for education and daycare. For a long time in the past, the intellectually-abnormal attended segregated special schools (the so-called ‘grammar schools’). The expansion of comprehensive education has of course reduced this segregation, but wealthy people still pay for their intellectually-abnormal children to attend residential special schools at places like Eton and Harrow. It is, however, after secondary school that non-inclusive policies dominate. Most intellectually abnormal at present go to segregated institutions (the so-called ‘universities’). Fortunately, changes in government policy has meant that these institutions have expanded to include a much wider range of students who are not intellectually abnormal. However, universities still have a major role providing sheltered daycare for many people with intellectual abnormalities for the whole of their adult lives. This is often presented as being ‘work’, but you will often find low levels of activity, most of which comprises pointless and unfulfilling activities like teaching and marking.
The situation is hardly better in the residential circumstances of intellectually-abnormal people. Many seek their friends exclusively among other members of this group, and even marry each other. They often live together in communal settings (particularly when they are ‘students’), and tend to congregate in particular areas of towns. The position is far worse in the USA, where there are rumoured to be whole college towns of people with intellectual abnormalities.
What should be done about this? Following the principles of Valuing People etc, we should enable people with intellectual abnormalities to choose the lifestyles of the general population. There is no reason why they can not work in open employment such as shopwork, clerical activities, and manual assembly-work. Indeed, thanks to recent Government economic policies, more and more intellectually-abnormal people are moving into such jobs after they leave university. We should emphasises the importance of intellectually-abnormal people taking part in ordinary community activities (like going to bingo, the pub and the dog track), which will help them develop friendships with ordinary people. In that way, we will move to a society where everyone will increasingly be the same.
Wednesday, 17 June 2009
How to drive the Worcestershire way
Speed limits. Always drive 10 miles/hour less than the local speed limit. This will avoid you being caught speeding if your speedometer is inaccurate.
Brakes. The brake pedal is the most useful part of the car. Push the brakes on when you are coming to a corner (any corner), when you see a car coming towards you on the other side of the road, when you are about to go up a hill, when you are going down a hill, and when you can see some traffic lights in the distance ahead of you.
Traffic lights. Wait a bit after the lights have changed to green. You never know if they might suddenly change back to red again.
Roundabouts. Never go on a roundabout if you can see another car approaching it. When you go on a large roundabout, always stick to the left-hand lane whether you are going left, straight ahead or right.
Country lanes. These do not have white lines down the middle, which means you can safely drive in the middle of the road. If you are lost, just stop the car and look at a map.
Motorways and other roads with three lanes each way. Very careful drivers avoid motorways altogether. But if you are on one, always drive in the middle lane. This gives you a clear view of the road, particularly important when other drivers try and pass you on the inside!
Filter lanes. Do not filter into traffic because you might then need to change lanes while moving. Instead, wait until the road is clear and drive straight over to the lane you want to be in. While you are waiting to do this, you should ignore the drivers behind you beeping their horns - they have obviously not learnt to appreciate how to drive the Worcestershire way.
Monday, 15 June 2009
Where have all the Marxists gone?
This is strange because capitalism is now truly in crisis, and there was a time in the recent past when Marxism seemed a powerful set of intellectual ideas. At the time I was active in the British Labour Party (in the 1970s and early 1980s), political debate among activists in my constituency party was dominated by different versions of Marxism. Various Trotskyist factions existed within the Party, of which the strongest was the ‘Militant Tendency’. This was essentially a self-contained and disciplined political party which pretended to be an informal group of voluntary newspaper sellers. There was a minestrone of other Marxists groups outside the Party, most of whom had names with various combinations of the words ‘Revolutionary’, ‘Communist’, ‘Socialist’ or ‘Workers’. None of these factions had much mass support: they were stage armies of would-be leaders of a revolutionary struggle which they (and many others on the Left and Right at that time) believed was imminent.
The lack of mass support did not seem important because British politics is dominated by secretive committees and caucuses. This gives great political leverage to small and well-organised groups. Control of constituency labour parties can be won by getting a few activists nominated from poorly-attended or inactive local branches, who can then nominate candidates for Parliament who can be elected on a Party ticket even though their voters usually have little idea of their personal opinions (or personal morality). Many of the most ambitious local politicians in the Labour Party in the early 1980s rose either as members of one of the Marxists factions, or by doing deals with them. Either way, this required sharing a political language of class struggle against capitalism. In reality, ‘class struggle’ amounted to organising local and Parliamentary election campaigns, supporting strikes, and passing long and angry resolutions to be despatched to the Party’s National Executive Committee.
This all made Marxist ‘entryism’ a successful strategy up to a point. Members of various factions could become councillors and a few became Members of Parliament. But once Marxists had made it onto the national stage, they showed an extraordinary ability to lose votes, particularly when they tried to use local councils to introduce ‘socialism in one borough’. The Party leadership realised at that point that they needed to act against the factions. But a more important cause of the decline of Marxism was the rapid disappearance of communism in Eastern Europe, and, more locally, the continued success of the Conservative Party in British elections. In other words, the struggle against capitalism seemed to have been lost, and ambitious politicians decided they must find another route to power. This other route involved following Tony Blair in portraying Labour as competent and non-threatening, but also being relaxed about major and growing inequalities in wealth, subservience to the USA in foreign policy, and the parcelling out public services to the great profit of an inter-connected web of management consultancy, IT and PFI companies.
This is a world away politically from the Marxism of the early 1980s, but the ex-Marxists who now dominate the Labour Party have retained many of their old habits of mind. There is still the bitter factionalism, the preference for secretive committee-room conspiracies, and the contempt for public and for democratic politics. Gordon Brown became Prime Minister as a result of the first two of these, while his failure to hold a general election shortly afterwards illustrates the third. What we miss from Marxism, and what has been abandoned by these politicians, is the political energy, the internationalism, and the dream of a society organised on a different and fairer basis.
Thursday, 4 June 2009
The old wandering track
Watkins’ ley-lines became adopted by ‘new age’ believers, the most rapidly-growing faith group in this country. Whereas Watkins proposed that ley-lines were just a means of finding your way across the wooded and boggy prehistoric landscape, new-agers speak of ley-lines leading along ‘lines of magnetic power’ across the ‘living rock’, and other such wonderful codswallop. I have a better explanation for lines of the landscape - to reach a good pint of ale. I have noticed that the pubs close to my home form a line, from the Fox at Lulsley, via the Talbot at Knightwick and the Admiral Rodney at Berrow Green, to the Crown at Martley. Extend the line Northwards, and you get to the Red Lion at Holt Health. Ale is an ancient drink, regarded as much safer to drink than water through much of English history. My ‘ale lines’ provide a much better motive for travelling than just following lines of magnetic power. Of course, if you stop at a few pubs on the way, you will no longer follow a straight path, which helps explain why the English landscape looks the way it does.
Wednesday, 3 June 2009
Dining in Yuppieland
My experience is that dining in these converted pubs is dispiriting. Earlier this year, I went with my mother to Henley-in-Arden, a beautiful small town along the Stratford Road in the heart of Yuppieland. The restaurant included mediaeval features, superb decor and friendly staff. The food was a work of art presented on elegant plates and looked wonderful. It was expensive but poorly-cooked. It somehow managed to be both tasteful and tasteless. This has been my experience with several dining expeditions to this part of the world. People in Yuppieland are prepared to spend a lot on their meals, like them to be an impressive visual experience in a fawning environment, but can not discriminate between good food and bad.
Not all of England is like this. To encounter good food made from fresh local ingredients, it is better to move further from the big cities. In my own area of rural Worcestershire, there are two places I visit which always delight. The first is the Talbot at Knightwick. This is a small hotel with its own brewery and farm. The food is imaginative, is sourced from the Teme Valley and is full of taste. The main draught beers are called ‘This’ and ‘That’, and these are the best beers I have ever tasted anywhere. The plates and beer glasses are not the products of designer studios, but who cares. The second place to eat is the Venture In in Ombersley. This is a small restaurant (with Michelin rosettes) owned by its cook, who lives with his family above the shop. I have a family connection with the Venture In because my son Andrew worked there as a kitchen porter one evening a week before going to University. The atmosphere in the kitchen was far from the kind of shouting match shown on television pictures of restaurants. Instead, there was an effective and friendly team of people dedicated to cooking in the best way from the best ingredients. The result is food that is as close to heaven as can be attained on earth.
Thursday, 28 May 2009
Where is the love?
How could this happen? One possible reason is staff burnout. Many people find talking to ill people emotionally taxing. This can wear down the kindest of people, who escape pressure by reducing the emotional content of their interaction with patients or clients, and convert their work to a set of technical procedures. But this raises the question of why organisations do not take steps to avoid burnout or why they continue to tolerate it among their staff. I think the main reason this happens is because the emotional content of professional and personal support work is seen as problematic: the direction of training this group of staff has therefore instead emphasised the acquisition of technical skills. Ethics is still taught as a subject in professional training, but has been reduced to a set of guidelines to follow in obtaining consent for research or for the application of medical, nursing or other procedures.
This loss of the emotional content of health and social care is part of a wider trend towards the excision of passion and feeling from organisational life. It is assumed that no-one works because they have skill and personal commitment, and a passion to apply it to their work. One consequence is the belief-in-practice (written into numerous guidelines and codes of practice and quality assurance manuals) that no-one can be trusted and no-one can perform well unless regulated and inspected or, if they are senior managers, given generous financial bonuses to do their work. This belief becomes self-fulfilling: staff become de-motivated and truculent under the weight of inspection, while senior managers become oriented solely to their bonuses and neglect wider responsibilities.
Of course it is often argued that such mechanisms of control are necessary to manage large organisations and complex centralised states. But this fails to ask whether we need such large organisations or whether our society should be so centralised. It is possible that this trend to centralisation has come about not because it is better at producing goods and services, but because it has generated a new and dehumanised ethos of organisational rationalism. This incorporates the distrust of human emotion and commitment, and hence the negative view of human nature. It proposes instead that humans are properly motivated only by a combination of financial rewards and penalties. This looks rational in the sense that economists speak of rational behaviour, but at its core are the darker emotions of greed, fear, and love of power.
Thursday, 14 May 2009
You are not included
‘Social exclusion’ is a wonderfully vague phrase promoted by governments because it avoids the need to talk about poverty and inequality. It is never made clear precisely which sorts of people are socially-excluded, or what precisely they are excluded from. Nor is evident why they are excluded in the first place. The term ‘socially-excluded’ has at different times been used to include people on low incomes, disabled people, people from ethnic and racial minorities, and even the elderly and the young. When the phrase first became part of political discourse in the 1990s, it was agreed that the problem lay with the excluders - with the lack of opportunities for well-paid work, with poor schools and with discrimination. Now governments increasingly suggest that the problem lies with the excluded, and that ‘social inclusion’ really means coming off welfare benefits. So the unemployed, single parents, the disabled and the chronically sick are to be hectored into taking poorly-paid work, irrespective of the fact that work of any kind is getting hard to find. Of course, if it all gets too much for the socially-excluded, cognitive behaviour therapy will be made available to cheer them up.
But the hectoring received by the ‘socially-excluded’ is only a rather more extreme version of what the rest of us have to put up with. This Government have sent leaflets to our homes telling us that we must all eat more healthily, that we must wash our hands, and use a tissue when we sneeze. Our children are assessed and examined more frequently than in any other country in the world, so that teachers can be harassed into improving their school’s position in national league tables. To supposedly prevent crime, we are watched, scanned and regulated by the largest CCTV network in the world. To supposedly prevent terrorism, basic legal protections are stripped away, with old men arrested as ‘terrorists’ if they heckle the prime minister at a Labour Party conference.
Yet there are small groups of people who have been exempt from all this monitoring, assessment and harassment - our financial sector and our members of Parliament. Financial leaders have squandered cash on foolhardy investments and driven their institutions to bankruptcy. Once found out, they have retired on very generous pensions while their banks have been funded at immense expense for the taxpayer. Many members of Parliament have used an elastic expenses system to refurbish one or more houses, make large capital gains, clean their moats, employ their spouses on generous salaries, and buy porn videos. When they are defeated at the next general election, they too will retire on a generous superannuation. Both bankers and members of Parliament explain away their behaviour by saying they acted within the rules, and that their activities were subject to audit. But this audit and inspection has been a sham - designed to give the impression of regulation while actually allowing them a free hand.
A sham if this kind exists because the powerful believe in their hearts that those who make the rules need not live by them. This network of beneficiaries constitute the ‘politically-included’, who live from the savings and taxes of the rest of us (the politically-excluded) who must pay for our own mortgages, rents, patio heaters, dogfood, moat-cleaning and so on. Nevertheless, this experience may be character-building, and we have a chance in future elections to improve the characters of our current politicians by enabling them too to experience the delights of political exclusion.
Wednesday, 6 May 2009
How to be Impatient - Part One
Few people seem able to praise impatience; cultural superiority lies with the stolid, the patient, and the inert. Never mind that most innovations have occurred because some impatient person got sick of waiting. This all means that impatient people, like me, are an oppressed cultural minority: we need to use our wits to survive. This obviously requires a lengthy book of hints, of the kind that get sold in airport bookshops. Regard this text as Part One.
The main area in which impatience is experienced, at least in England, is on the roads. I drive a small sports car, and enjoy zipping along the open road. However, a more usual traffic experience is to be at the end of a queue of slower vehicles, tailing along behind a Nissan Micra (the underpowered car for underpowered people). This shows the terrible paradox of driving - that the slower the car, the more likely it is to be in the front of the traffic with an unimpeded view of the road ahead.
So there is not much advice I can give impatient drivers. However, I do have a handy hint regarding another area of impatience: which queue to join in the supermarket check-out. My advice, based on sustained observation and testing, is to choose the queue with the fewest women. There are two main reasons:
1. Women are nicer than men. Women are more prepared than men to regard the staff on the checkout as people, worthy of conversation. This takes time. They will also attempt to help the checkout staff by giving the exact change for their purchases. This involves carefully sorting through their money and counting it out. This also takes time, and contrasts with the male approach of dropping a banknote.
2. Women are more careful with their money than men. Most men keep their cash in their trouser or jacket pockets. Women, on the other hand, keep their cash in a series of containers, each inside the other. The purse (designed to be small and difficult to get fingers into) is kept inside a handbag, which is often inside a shopping bag. Any purchase is therefore preceded by a sequence of money-discovery, and followed by a reverse sequence of money-concealment.
To come later: hints on how to break the buffet queue.
Monday, 4 May 2009
The conspiracy continued...
SIR – It is highly worrying that on Tuesday morning at Birmingham airport, 400 passengers were allowed to leave a flight from Brazil and go their separate ways – without any checks, records or advice.
Why did nobody go on the plane before embarkation, check the passengers and give advice?
It is now quite obvious that the authorities are not serious about containing the spread of this virus – and for very good reasons. The benefits of a worldwide killer pandemic are huge!
A large global loss of life through ostensibly “natural causes” would lift a huge burden from politicians, by reducing world over-population at a stroke without the enormous cost of any war. It also instantly reduces the demand on our stretched food stocks.
I would go further and say that it might have been surreptitiously and deliberately introduced and Mexico would be the ideal propagation nation. The pieces of the puzzle fit rather too neatly for my liking.
So, always use a hanky and wash your hands frequently.
STANLEY D PARR,Pershore
As I suspected, some people will find a conspiracy everywhere.
Sunday, 3 May 2009
The Great Mexican Swine Flu Conspiracy - Gate
What makes a good conspiracy theory? There should be a major disaster involving loss of life. Obvious natural events, like earthquakes, floods, and volcanoes must not be involved (although this will not stop enthusiastic religious believers from attributing these causes to the desire of a just and loving God to indiscriminately wipe out large numbers of guilty and innocent people as a sign of man’s fall from grace). There must be some real or imagined discrepancy in the official explanation for the disaster, which can ascribed to a desire by those responsible for the conspiracy to hide their involvement. Further attempts by governments to clarify causes of the disaster, hold commissions of enquiry etc, will be portrayed as a ‘cover-up’, usually involving the suffix ‘gate’. The fact that an effective cover-up would require complicity by the entire government, press and civil service becomes ‘evidence’ that it is a truly immense conspiracy.
What are the conspiracy theories of the future? I propose that Mexican Swine Fever could do the job. This began with predictions of pandemics and millions of fatalities. Then, all of a sudden, we heard of a few non-fatal infections. A conspiracy theorist would argue that the millions of deaths have and actually are taking place, but that governments are covering this up by taking over isolated warehouses and incinerating bodies in secret. Lots of people disappear every year, and this explains what is happening to them. Other bodies are being loaded on to ships and sent to the Gulf of Aden, where Somali pirates, who are part of the conspiracy, capture and sink the ships. We could imagine an even more comprehensive conspiracy theory, in which governments are responsible for the Swine Fever in the first place. This could be a result of using pigs for germ warfare experiments, to be tried out first on the Mexican population. Perhaps governments are in league with the Green Movement, who believe that only a massive reduction in the world’s population can - er - save the world’s population. Of course, the conspiracy theorists could themselves be part of an even wider conspiracy...
Thursday, 30 April 2009
Looking down on others' needs
But this task raises a real problem: on what basis can you assess the needs of other people? One traditional solution is to assume that all people have the same needs and should live their lives in the same way (usually involving the performance by the entire community of the same religious or patriotic rituals). But in more individualistic societies, another way of assessing needs is required. Fortunately, psychologists and social researchers have the answer: detailed lists of human needs and ‘objective’ measures of the quality of life. The oldest of these categorisations of needs, and the one most people have heard of, is Mazlow’s Hierarchy.
Mazlow proposed that human beings seek to satisfy needs in a set progression, starting from the most basic physiological needs. Once these needs are satisfied, human beings can aim to satisfy successively higher needs, leading ultimately to self-actualisation and self-transcendence. What do these last two phrases mean? ‘Self-actualisation’ for Mazlow meant reaching one’s full potential, which he believed was marked by creativity, an internalised morality, spontaneity, and closeness to others. ‘Self-transcendence’ referred to peak experiences in which a person experiences a sense of spiritual fulfilment.
There are some obvious problems with Mazlow’s approach. People are too diverse to fulfill their needs in the orderly way he suggested. We can all think of examples of people who follow a different hierarchy: people who starve themselves for beauty or for religious belief, and people who seek out risk and danger. People also differ in what they regard as ‘self-transcendence’. People like Aldous Huxley and Timothy Leary proposed it could be achieved through drugs (so for that matter did Ian Drury in his song Sex and Drugs and Rock and Roll). To most of us, however, drug-taking looks more like self-destructive hedonism. I suspect that Mazlow esteemed ‘self-transcendence’ because he was a humanist psychologist. If models wrote hierarchies of needs, they would probably place beauty and poise at the top, while footballers would place agility and teamwork. In other words, setting up a supposed hierarchy of needs becomes an implicit way of ranking other people or looking down on their needs. This would not matter too much, except that hierarchies of needs get imposed on others.
The most imposed hierarchy of all is that of rationality. Many philosophers (who as a rule are rather good at reasoning) have decided that rationality is not only at the top of the hierarchy but also the true marker of humanity. People can then be ranked according to their reasoning powers, with questions raised about the extent to which people with severely-impaired reasoning are really human. This has led to the ‘thought experiment’ by the philosopher Peter Singer and others which compares the relative worth of a severely-disabled person and one of the more intelligent animals. Singer’s views are subtle and have been misrepresented, but anxiety about this kind of though experiment are understandable. They are based on the fear that somewhere in a garret is a lonely soul planning a rise to political power in which thought experiments will be implemented. Then the killings will begin again.
Monday, 20 April 2009
My life as a steam engine
When I was young, absolutely no-one reported that they were experiencing ‘stress’. The common metaphor of the time was the steam engine, which still dominated rail transport. Like a steam engine, people said they were ‘under pressure’ when life became difficult, the answer to which was ‘to let off steam’. This involved some physical activity or other form of release such as drinking and socialising. By the 1970s, however, steam engines and rail transport had largely been replaced in people’s minds by motor transport, and people came to regard themselves as being a type of car. Cars and other fast-moving machines suffer from metal fatigue and stress, and this became the dominant metaphor for the emotional state experienced when people face adversity. The conventional answer to being ‘under stress’ or ‘stressed out’ is inactivity, or possibly handing over responsibility for their condition to a therapist. There is no shortage of these, all promoting their services as helping people avoid the sad fact that life includes a fair share of loss, pain and grief.
As for me, I still think of myself as more of a steam engine than a car. When life gets difficult, I prefer to become more active. Of course, it is still helpful to unburden myself on my wife and others, even if it does make them feel stressed.
Sunday, 19 April 2009
Synthetic nostalgia
The origin lies with my parents, particularly my father. He was born in 1914 in what seems to have been a reasonably prosperous family which lived in villages and suburbs on the outskirts of Birmingham. He told me that, when he was a child, the family had a domestic servant. Yet his parents were defrauded of their business, and he came home from school one day to find the family’s furniture and other possessions on the street following eviction. He passed the exams for the local grammar school, but failed the interview. In the absence of a secondary education, he did a variety of poorly-paid jobs, including truck driver (at the age of 14!), chauffeur, workhouse clerk, and even coalman. Eventually, he became a welder on the Land Rover track in Solihull and then in Birmingham. I remember him cycling to work in the rain, and the burn marks down his chest from welding sparks. I am not sure how these experiences shaped him, but my father always had a strong sense of social justice and a commitment to trade unions and the Labour Party.
I spent the early years of my life following these footsteps and inheriting his ambitions. I was active in Labour Party politics and imagined myself as a future MP. I felt part of a working-class movement that would make society more just and treat all people with respect. So when I first visited the Black Country Museum, I identified with the past lives of the working-class, and adopted this synthetic nostalgia. I do not believe my experience is unusual: the difference between us all is the nostalgia we adopt. I have a genial colleague at work who has adopted the mannerisms of an Oxbridge scholar even though he comes from near Bolton. I know of people of marginal religious faith who have adopted the lifestyle and clothing of the committed believer to assert their membership of a lost identity.
My father, Dad, died in 2000. He spent the last 15 years of life unable to speak following an operation for cancer of the throat. As a man who enjoyed talk and argument, this must have been a great loss, but he managed his life with day-to-day stoic courage. I do not pass a day without thinking of him. From both parents, I learnt the importance of education as a means of advancing from dreary low-income work, and have spent my working life in fulfilling occupations with few money worries. I still retain my father’s political convictions, though, unlike him, I feel unrepresented. The political party that he and I worked for year after year has all but disappeared. Its successor, which calls itself the ‘New Labour Party’, has betrayed almost all the hopes of those who voted it into office. After the Iraq War began, I tore up my party card. I am no longer nostalgic for an imagined working-class past, but my anger and sense of betrayal remains strong.