the poster for Krzysztof Kieslowski’s A Short Film About Killing (1988)
I don’t believe in the death penalty. In this, I’m in the majority, globally. I’m not sure when exactly I became against it; until at least the age of 12 I was pretty much a proto-fascist with an ‘eye-for-an-eye’ sense of justice, as boys tended to be in those days and for all I know still are. But I know that by the time I saw Krzysztof Kieslowski’s brilliantly grim A Short Film About Killing (Krótki film o zabijaniu) when I was 16 or so I was already anti-death penalty and have remained so ever since.
My reasons are, typically, kind of pedantic. There are many obvious arguments against it; there’s the ‘what if you accidentally kill the wrong person’ argument and that’s a pretty strong one – it has happened and does happen and is irreversible. There’s the fact that the death penalty seems to have a negligible effect on the crime rate. In fact, countries with the death penalty on the whole seem to have more rather than less murders (not that there’s necessarily a link between those two things). Even from the coldest and most reptilian, utilitarian point of view of getting rid of the problem of prison overcrowding, any possible benefit is negated by the fact that in most countries with the death penalty, prisoners spend years on death row being fed and housed, rather than being quickly and efficiently ‘processed.’ There’s also the Gandalfian(!) argument from The Lord of the Rings; “Many that live deserve death. And some that die deserve life. Can you give it to them? Then do not be too eager to deal out death in judgement. For even the very wise cannot see all ends.” This wasn’t just a handy deus ex machina because Tolkien needed Gollum to survive in order to destroy the ring. It was that, but Tolkien was also a devout and serious Christian and that was his moral outlook. Thank the gods that unlike his friend CS Lewis, he deliberately left religion out of his books though! In the Biblical commandment Thou Shalt Not Kill, the Christian/Jewish god doesn’t list any exceptions or mitigating circumstances – in that one instance. Of course elsewhere in the Bible there are many circumstances where humans killing humans is considered appropriate and even righteous – the ultimate irony being that Jesus, kind of like an anti-Gollum, has to suffer death through violence to achieve his purpose. Religion is odd; but I’m not a Christian or Jew.
All of those points are relevant, but for me personally, it’s far simpler than that; if you can be legally killed, that means that in the eyes of the state there’s essentially nothing wrong with killing people. I think there is, and I don’t think that it should just be a matter of having the right paperwork. In essence, to kill a murderer is not telling them ‘what you did is wrong‘ so much as ‘you did it wrong‘ which I don’t think is a minor difference. And on top of that, there’s the whole question of who you are handing this responsibility of life and death to. I have a lot of respect for some lawyers, attorneys, judges, police officers etc, but there are others that I wouldn’t trust with my lunch, let alone my (or anyone else’s) life. States have a character, and often it is institutionally biased regarding race, class, gender and sexuality. Giving that kind of power within that kind of framework seems likely to make far more problems than it solves. But even in non-death-penalty countries like the UK we routinely give people the legal right to take other people’s lives, all they have to do is join the armed forces.
British volunteers in the International Brigade, 1937
I’m no more consistent than anyone else and my attitudes have their exceptions and contradictions. I (predictably) don’t philosophically differentiate between the military and mercenaries, because what ‘serving your country’ means in practical terms is carrying out whatever the policy of your government is that week, with no certainty that it won’t be contradicted by a new policy (or a new government) the next week and if enemies suddenly turn out to be allies or vice versa, the dead remain dead. That said – here’s the contradiction – I’m not a pacifist absolutist either, and I think, or like to think that if an invading army arrived in my country I’d take arms against it. These things are particular though; everyone likes to think they’d fight for a good cause, but the Spanish Civil War stands out for the number of anti-fascist fighters from all over the world who took up arms in defence of Spain. But that happened partly because so many people were ready to – and wanted to fight. Many of those – George Orwell is a prominent and typical example – belonged to the generation who had been just too young to fight in World War One and whose feelings about war – including a considerable amount of survivor’s guilt – had been shaped by it. And the fascist attack on the Spanish republic gave them a clear-cut situation to intervene in, in a way that the more political rise of fascism in Italy and Germany didn’t.
But anyway, the death penalty. People of course do terrible things, but although lots of them are significantly more horrific than a lethal injection or the electric chair, the end result is the same. Being – odd, brief segue but bear with me, it’s relevant – a fan of black metal music, the subject of death and murder is one you come across in a different way from just being, say, a fan of horror movies. Because the poser-ish ‘darkness’ of black metal spills over (though less than it used to) into ‘real life,’ almost as if the kind of art you make bears some relation to the kind of person you are. I won’t go into the tedious-but-fascinating Lords of Chaos stuff about Mayhem & Burzum or Absurd because it’s not quite relevant here, but the story of Smutak (Pavel Selyun) who ran Morak Production record label in Belarus is.
In 2012 Selyun discovered that his wife, the artist and singer Frozendark (Victoria Selyunova) was having an affair with the artist, zine editor and musician Kronum (Alexey Vladimirovich Utokva). Sticking with the psuedonyms seems appropriate, so anyway; Smutak murdered both Frozendark and Kronum, dismembered them and was apprehended on the Subway three days later with Kronum’s head (or skull; same difference I suppose – some accounts say he boiled the head – I don’t need to know) in a bag. After his arrest, he was imprisoned in Minsk and after a confession gained under torture and the failure of various appeals he was executed two years later, by being shot in the back of the head. A horrible postscript that demonstrates how the death penalty punishes the innocent as well as the guilty; after the execution the authorities failed to hand over Smutak’s body to his mother or tell her where he’s buried, the case was handed to the UN Court of Human Rights.
Not many people (and certainly not me) would say that Selyun didn’t ‘deserve’ his treatment. But still. He possibly tortured and definitely killed people and then was tortured and killed. There is a kind of balance there, but it’s one in which the act of torturing and killing itself is made neutral. Whoever tortured and killed Smutak doesn’t need any kind of defence because they did it in the name of the law, but the idea that torturing and killing is morally neutral because you don’t have any emotional investment in the act is an odd one. Smutak had nothing to gain from his actions other than some kind of horrible satisfaction. The person or people who did the same to him got paid for it. Which is morally, what? Better? He reportedly felt the same kind of fear as his victims; well good, I guess, but that did nothing to benefit the victims. It may have pleased the victims’ relatives but I wouldn’t want to examine that kind of pleasure too closely.
The current case of Luigi Mangione is far stranger. It’s the only time I can recall that the supporters (in this case I think ‘fans’ would be just as correct a word) of someone accused of murder want the suspect to be guilty rather than innocent. Whether they would still feel that way if he looked different or had a history of violent crime or had a different kind of political agenda is endlessly debatable, but irrelevant. It looks as if the State will be seeking the death penalty for him and for all the reasons listed above I think that’s wrong. But assuming that he’s guilty, which obviously one shouldn’t do (and if he isn’t, Jesus Christ, good luck getting a fair trial!) Mangione himself and some of his fans, should really be okay with it. If he is guilty, he hasn’t done anything to help a single person to get access to healthcare or improve the healthcare system or even effectively protested against it in a way that people with political power can positively react to. UnitedHealthcare still has a CEO, still has dubious political connections and still treats people very badly. That doesn’t mean that it’s an unassailable monolith that can never be changed, but clearly removing one figurehead isn’t how it can be done.
But more to the point; why does the killer (assuming their motives are the ones that are being extrapolated from the crime) care anyway? If actually shooting someone dead in the street is okay, then surely being indirectly responsible for the misery and possible deaths of others is barely even a misdemeanour. It amounts to the kind of Travis Bickle movie logic I’m sure I’ve sneered about elsewhere; complaining about the decay of social values and then committing murder is not reducing the sum total of social decay, it’s adding to it. A society where evil CEOs are shot dead in the street is a society where human beings are shot dead in the streets and that becoming acceptable is not likely to be the pathway to a more just, equal or happy society.
Michael Haneke’s disturbing Benny’s Video (1992)
What the death penalty does do, and probably a key part of why it’s still used in some countries, is offer a punishment that seems (in the case of murder at least) to fit the crime. Interestingly, public executions – which counterintuitively seem to have no better track record as a deterrent than any other kind – are now vanishingly rare. Part of that is no doubt to do with public disgust and part with institutional secrecy and shame, but I imagine that part of it is also the fear that the public would enjoy it too much. I’m not sure if I would think that if it wasn’t for the spate of Islamic State beheadings that were so widely watched on the internet back in the early 2010s (was it?) I watched one, like most people seem to have, and still wish I hadn’t; but you can’t un-ring a bell. That was at the back of my mind when I wrote about saints and martyrdom for this site and I can bring images of it to mind horribly easily. But even before that it shouldn’t have surprised me – like many other teenage horror movie fans in the pre-internet era I watched exploitation videos like Face of Death that featured executions, accidents etc, and in doing so realised that I was a horror fan and not whatever fans of that are. I should have learned my lesson there, but it’s undeniable that these things have a murky kind of fascination; since then, thanks to one of my favourite writers, Georges Bataille, I’ve ended up reading about Lingchi (‘Death by a Thousand Cuts’) and looking at the chilling and depressing photos of it, been appalled by postcards of lynchings, seen revolting photographs of soldiers’ desecrated bodies and murder victims… I haven’t gotten used to those images and I hope I never will. Teenage me would no doubt sneer at that because he thought that things that are ‘dark’ are cool, but that seems like a laughable and childish attitude to me now, so I can take his sneering. I seem to be edging towards the point that Michael Haneke is making in Funny Games (1997), which I find a bit tiresome and preachy (even more so the remake), but I’m not. I disagree with the premise of that film because I do think there’s a difference between fictional horror and real horror, and that enjoying one isn’t the same as enjoying the other. I think his 1992 film Benny’s Video makes a similar but much more subtle and complex point far better.
Imprisonment (whatever your views on the justice system) is a pretty unsatisfactory solution for most crimes, but it’s difficult to think of a better one which doesn’t essentially exonerate the kind of behaviour we want to characterise as abnormal or criminal. Stealing from a thief is obviously ‘justice’ in the eye-for-an-eye sense, but as a punishment it’s laughable. Raping a rapist would be grotesque and double the number of rapists in the room every time it happened. But even so, it’s never going to be comfortable that the tax payer is contributing to the relative comfort of someone like (I’ll only mention dead ones, this isn’t a complaint about the legal system being soft on psychopaths) Fred West. A solution l think I might suggest is one which I’m very dubious about myself from lots of different humanitarian, psychological and philosophical points of view; why not offer (and that word alone would make people angry) ‘monsters’ – the kind of killers in a category of their own, who admit to horrendous acts of murder and torture and whose guilt is not in doubt – those who will never be allowed freedom – the choice of a lethal injection rather than life imprisonment? That’s a horrible thing to contemplate, but then so is paying for the meals and upkeep of someone like Ian Brady, especially when he essentially had the last laugh, exercising his little bit of power over the families of his victims and having his self-aggrandising bullshit book The Gates of Janus published.
Anyway, that last part was kind of icky and uncomfortable, but so it should be – the whole subject is. So for what it’s worth, those are my thoughts on the death penalty. Time for a shower; until next time, don’t murder anyone please.
eh, Mel Gibson: but he played a good Hamlet (dir Franco Zeffirelli, 1990)
With apologies to Marcel Proust – but not very vehement apologies, because it’s true – the taste of honey on toast is as powerfully evocative and intensely transporting to me as anything that I can think of. The lips and tongue that made that association happen don’t exist anymore and neither does the face, neither do the eyes, and neither does one of the two brains and/or hearts* that I suppose really made it happen (mine are still there, though). In 21st century Britain, it’s more likely than not that even her bones don’t exist anymore, which makes the traditional preoccupation with returning to dust feel apt and more immediate and (thankfully?) reduces the kind of corpse-fetishising morbidity that seems to have appealed so much to playgoers in the Elizabethan/Jacobean era.
Death & Youth (c.1480-90) by the unknown German artist known as The Master of the Housebook
Thou shell of death, Once the bright face of my betrothed lady, When life and beauty naturally fill’d out These ragged imperfections, When two heaven-pointed diamonds were set In those unsightly rings: then ’twas a face So far beyond the artificial shine Of any woman’s bought complexion (The Revenger’s Tragedy (1606/7) by Thomas Middleton and/or Cyril Tourneur, Act one, Scene one)
*is the heart in the brain? In one sense obviously not, in another maybe, but the sensations associated with the heart seem often to happen somewhere around the stomach; or is that just me?
More to the point, “here hung those lips that I have kissed I know not how oft“, etc. All of which is beautiful; but for better or worse, a pile of ash isn’t likely to engender the same kind of thoughts or words as Yorick’s – or anybody’s – skull. But anyway, the non-existence of a person – or, even more abstractly, the non-existence of skin that has touched your skin (though technically of course all of the skin involved in those kisses has long since disappeared into dust and been replaced anyway) is an absence that’s strange and dismal to think about. But then most things don’t exist.
Vanitas: Still Life with Skull (c.1671) by an unknown English painter
But honey does exist of course; and the association between human beings and sugary bee vomit goes back probably as long as human beings themselves. There are Mesolithic cave paintings, 8000 years old or more, made by people who don’t exist, depicting people who may never have existed except as drawings, or may have once existed but don’t anymore, plundering beehives for honey. Honey was used by the ancient Egyptians, who no longer exist, in some of their most solemn rites, it had sacred significance for the ancient Greeks, who no longer exist, it was used in medicine in India and China, which do exist now but technically didn’t then, by people who don’t, now. Mohammed recommended it for its healing properties; it’s a symbol of abundance in the Bible and it’s special enough to be kosher despite being the product of unclean insects. It’s one of the five elixirs of Hinduism, Buddha was brought honey by a monkey that no longer exists. The Vikings ate it and used it for medicine too. Honey was the basis of mead, the drink of the Celts who sometimes referred to the island of Britain as the Isle of Honey.
probably my favourite Jesus & Mary Chain song: Just Like Honey (1985)
And so on and on, into modern times. But also (those Elizabethan-Jacobeans again) “The sweetest honey is loathsome in its own deliciousness. And in the taste destroys the appetite.” (William Shakespeare, Romeo and Juliet (c.1595) Act 2, scene 6) “Your comfortable words are like honey. They relish well in your mouth that’s whole; but in mine that’s wounded they go down as if the sting of the bee were in them.”(John Webster, The White Devil (1612), Act 3. Sc.ene 3). See also “honey trap”. “Man produces evil as a bee produces honey.”You catch more flies with honey.
But on the whole, the sweetness of honey is not and has never been sinister. A Taste of Honey, Tupelo Honey, “Wild Honey,” “Honey Pie”, “Just like Honey,” “Me in Honey,” “Put some sugar on it honey,” Pablo Honey, “Honey I Sure Miss You.” Honey to the B. “Honey” is one of the sweetest (yep) of endearments that people use with each other. Winnie-the-Pooh and Bamse covet it. Honey and toast tasted in a kiss at the age of 14 is, in the history of the world, a tiny and trivial thing, but it’s enough to resonate throughout a life, just as honey has resonated through the world’s human cultures. Honey’s Dead. But the mouth that tasted so sweetly of honey doesn’t exist anymore. Which is sad, because loss is sad. But how sad? Most things never exist and even most things that have existed don’t exist now, so maybe the fact that it has existed is enough.
“Most things don’t exist” seems patently untrue: for a thing to be ‘a thing’ it must have some kind of existence, surely? And yet, even leaving aside things and people that no longer exist, we are vastly outnumbered by the things that have never existed, from the profound to the trivial. Profound, well even avoiding offending people and their beliefs, probably few people would now say that Zeus and his extended family are really living in a real Olympus. Trivially, 70-plus years on from the great age of the automobile, flying cars as imagined by generations of children, as depicted in books and films, are still stubbornly absent from the skies above our roads. The idea of them exists, but even if – headache-inducing notion – it exists as a specific idea (“the idea of a flying car”), rather than just within the general realm of “ideas,” an idea is an idea, a thing perhaps but not the thing that it is about. Is a specific person’s memory of another person a particular thing because it relates to a particular person, or does it exist only under the larger and more various banner of “memories”? Either way, it’s immaterial, because even though the human imagination is a thing that definitely exists, the idea of a flying car is no more a flying car than Leonardo da Vinci’s drawing of a flying machine was a flying machine or that my memory of honey-and-toast kisses is a honey-and-toast kiss.
If you or I picture a human being with electric blue skin, we can imagine it and if we have the talent we can draw it, someone could depict it in a film, but it wouldn’t be the thing itself, because human beings with electric blue skin, like space dolphins, personal teleportation devices, seas of blood, winged horses, articulate sentient faeces and successful alchemical experiments, don’t exist. And depending on the range of your imagination (looking at that list mine seems a bit limited), you could think of infinite numbers of things that don’t exist. There are also, presumably, untold numbers of things that do exist but that we personally don’t know about or that we as a species don’t know about yet. But even if it was possible to make a complete list of all of the things in existence (or things in existence to date; new things are invented or develop or evolve all the time), it would always be possible to think of even more things that don’t exist, – simply, in the least imaginative way, by naming variations on, or parodies of everything that does exist. So supermassive black holes exist? Okay, but what about supertiny pink holes? What about supermedium beige holes? This June, a new snake (disappointingly named Ovophis jenkinsi) was discovered. But what about a version of Ovophis jenkinsi that sings in Spanish or has paper bones or smells like Madonna? They don’t exist.
JAMC Honey’s Dead, 1992
Kind of a creepy segue if you think about it (so please don’t), but like those beautifully-shaped lips that tasted of honey, my mother no longer exists, except as a memory, or lots of different memories, belonging to lots of different people. Presumably she exists in lots of memories as lots of different people who happen to have the same name. But unlike supermedium beige holes, the non-existence of previously-existing things and people is complex, because of the different perspectives they are remembered from. But regardless, they are still fundamentally not things anymore. But even with the ever-growing, almost-infinite number of things, there are, demonstrably, more things that don’t exist. And, without wishing to be horribly negative or repeating things I’ve written before, one of the surprises with the death of a close relative was to find that death does exist. Well, obviously, everyone knows that – but not just as an ending or as the absence of life, as was always known, but as an active, grim-reaper-like force of its own. For me, the evidence for that – which I’m sure could be explained scientifically by a medical professional – is the cold that I mentioned in the previous article. Holding a hand that gets cold seems pretty normal; warmth ebbing away as life ebb away; that’s logical and natural. But this wasn’t the expected (to me) cooling down of a warm thing to room temperature, like the un-drunk cups of tea which day after day were brought and cooled down because the person they were brought for didn’t really want them anymore, just the idea of them. That cooling felt natural, as did the warming of the glass of water that sat un-drunk at the bedside because the person it was for could no longer hold or see it. That water had been cold but had warmed up to room temperature, but the cold in the hand wasn’t just a settling in line with ambient conditions. It was active cold; hands chilling and then radiating cold in quite an intense way, a coldness that dropped far below room temperature. I mentioned it to a doctor during a brief, unbelievably welcome break to get some air, and she said “Yes, she doesn’t have long left.” Within a few days I wished I’d asked for an explanation of where that cold was coming from; where is it generated? Which organ in the human body can generate cold so quickly and intensely? Does it do it in any other situations? And if not, why not? So, although death can seem abstract, in the same sense that ‘life’ seems abstract, being big and pervasive, death definitely exists. But as what? Don’t know; not a single entity, since it’s incipient in everyone, coded into our DNA: but that coding has nothing to do with getting hit by cars or drowning or being shot, does it? So, a big question mark to that. Keats would say not to question it, just to enjoy the mystery. Well alright then.
Klaus Nomi as “the Cold Genius” from his 1981 version of Purcell’s “The Cold Song”
But since most things *don’t* exist, but death definitely does exist, existence is, in universal terms, rare enough to be something like winning the lottery. But like winning the lottery, existence in itself is not any kind of guarantee of happiness or satisfaction or even honey-and-toast kisses; but it at least offers the possibility of those things, whereas non-existence doesn’t offer anything, not even peace, which has to be experienced to exist. We have all not existed before and we will all not exist again; but honey will still be here, for as long as bees are at least. I don’t know if that’s comforting or not. But if you’re reading this – and I’m definitely writing it – we do currently exist, so try enjoy your lottery win, innit.
Something silly about music next time I think.
Ancient Roman vanitas mosaic showing a skull and the wheel of fortune
But superstition, like belief, must die,
And what remains when disbelief has gone?
Grass, weedy pavement, brambles, buttress, sky,
A shape less recognisable each week,
A purpose more obscure.
Philip Larkin, Church Going (1954)
Grass, weedy pavement, brambles, buttress, sky
Given that Christianity seemed to be – in the sense of being a kind of shared societal glue – on its way out in the 1950s, and was undermined further by the social revolutions of the 1960s and 70s, it’s surprising in a way that churches are still standing at all. But what Larkin, for all of his humanist cynicism didn’t foresee, is what seems the obvious fate of churches in the 21st century: they won’t be allowed to peacefully moulder into dust and neglect like the menhirs and cairns of previous eras – they get sold instead.
At the time of writing, the Church of Scotland has seventeen churches for sale, among other kinds of properties and plots of land. The same thing isn’t happening with Catholic churches, Mosques, Synagogues or Mormon temples or whatever it is that Scientologists have – not because of anything inherently superior about those religions or the quality of their followers’ faith, but because, at the time when these churches were built – mostly in the 19th century, but some even earlier – the Church of Scotland was something you had to opt out of, not something you had to join. And therefore, in a way – although not of course a legally binding way – the Church of Scotland is selling off something which belongs to the people of Scotland.
The idea that money is more important to the Church of Scotland than the buildings that were at the centre of the spiritual and social lives of generations of people (and also, the place that God lived, I guess) seems grotesque, but there it is. It’s just bricks (or stone) and mortar, after all; or that, presumably is the logic, because God doesn’t actually live in a stone building but in either heaven or the hearts of believers etc, etc. And yet, if it’s just a building, how come people can only vandalise houses or schools or barns, but they can “desecrate” churches? “De-consecration” – what the church does in order to render its buildings saleable – is just a non-inflammatory way of saying desecration. De-consecrating the church doesn’t affect the material of the building, but it does remove its purpose – but what it can’t do is remove its history. So if you buy a church, what is it that are you actually buying? In a book I liked as a teenager, Terry Brooks’s Magic Kingdom For Sale – Sold! (1986), a depressed lawyer called Ben Holiday buys what turns out to be something like Narnia or Middle Earth, from a catalogue (nowadays it would be from a website). If Mr Holiday bought a church, he wouldn’t be mystically transported to an otherworldly realm, but he would – and the buyers of these buildings do – become the owner of a place where thousands of people were, in a meaningful way, transported to a place where, whatever the privations and terrors of their daily lives might be, things made some sort of black-and-white sense. Somewhere that virtue was rewarded with eternal paradise, vice was punished with eternal damnation and the person in the pulpit had the correct answers to whatever questions life was throwing at you. You don’t have to believe in any of that to realise that it was (and to some extent I suppose still is) important.
Like, I’m sure, many convinced lifelong atheists (and I’m a very un-spiritual one at that), I love churches. The architecture, the fixtures and fittings, the solemn atmosphere. The idea of building on top of (Native American) Indian burial grounds was enough to fuel horror fiction and urban legend for a century; will turning churches into houses, flats and offices do something similar? Probably not; although some of the churches for sale do indeed still have graveyards attached, the churches themselves, whether used or not, are utterly familiar to the local people. Like the Indian burial grounds, they have, for these people, always been there, but unlike them, they have always been visible, and have far more mundane connotations. They aren’t, or weren’t just the places people got married or had funeral services, they are places where, very recently, a few times a year you trooped along with your primary school classmates to hear about the less commercial, less fun aspects of Easter or Christmas and to sing a few hymns. In short, even now churches aren’t, or are rarely “other” in the way that (to non-indigenous settlers and their descendants) Indian burial grounds are. But, after generations will they still be familiar in that way, or will they become just funny-shaped houses? Who knows, but it’s sad to think so.
However much one does or doesn’t believe in the mythology that put them there, churches, just as local landmarks, bear the weight of memory, just as schools, war memorials, statues and monuments do. Although a valuable and significant thing, it’s a personal, private and unknowable kind of value; nostalgia, in its original, Greek meaning of ‘homecoming pain’ can be evoked in all of its intense complexity by almost anything, and in your own private iconography a road sign or piece of weed-strewn wasteland is likely to be as potent as, or even more potent than the more obvious celestial symbolism of the heaven-pointed steeple and arched windows. But the fact remains that the hopes, beliefs, dreams, grief and pain of generations was directed towards the church like lightning towards the weather vane that surmounts it; there is a kind of power just in that.
So what should be done with churches? You can’t keep everything forever, after all, and the Church of Scotland is, strange though it is to say it, a business. The people used to belong to it; it never belonged to the people and its churches are not public property in anything but the spiritual sense. But perhaps they should be: granted, they only reflect one strand of what is now a multicultural (and what was always a multi-faith) nation, but it’s a strand that informs attitudes and ways of life that contribute, both negatively and positively to the character of the country and its culture to this day. And although I don’t personally believe in the idea that buildings and land absorb a kind of psychic residue that manifests itself in the ghosts, hauntings and folklore beloved of digital TV channels, I feel like they should.
Those fundamental life events; christenings, marriages, funerals, wars, disasters – all of those lost people and all of that vanished emotion, should have some kind of monument or repository – and what better place than a church? Still; maintaining empty buildings purely for the sake of their history is an expensive, ethically dubious business and hardly an indicator of cultural good health. Finding new uses for these kinds of buildings that somehow respects their history is no easy task either. Personally, I’d like the government to buy them and use them to display the large percentage of publicly owned art that is currently languishing in the storerooms of galleries and museums, fulfilling in some ways at least, the National Galleries of Scotland’s strategic plan: “we will make the national collection accessible to all and inspire curiosity across the world. We want to connect with our audiences and with each other in new, collaborative and involving ways.” It would be appropriate in a way; human beings create art as god is supposed to have created people after all, and people with or without gods make art; it expresses many of the same fundamental impulses and emotions as religion. But it’s hardly an idea that’s likely to capture the public imagination, except in the negative sense that ambitious government spending on the arts – not that there has been much of that – always invites manufactured outrage. Ah well, it’s probably best to just make them into flats.
* I realise the double meaning was already implied in the title of Larkin’s poem, but why not render it completely unsubtle with a comma?
ceramic sculpture of a Moon Goddess and her rabbit or hare partner, Mexico, c.700 AD
Imagine a culture so centred on wealth, property and power that it becomes scared of something as fundamental to human existence as sex, and frets endlessly about what it sees as the misuses of sex. A culture that identifies breeding so closely with with money, wealth and status, and women so closely with breeding and therefore with sex that, when looking to replace the traditional symbols of birth and regeneration it rejects sex and even nature and, in the end makes the embodiment of motherhood a virgin and the embodiment of rebirth a dead man. Unhealthy, you might think; misanthropic even – and yet here we are.
But when that misanthropic culture loses the religious imperative that fuelled it for centuries, what should be waiting but those ancient symbols of fertility; rabbits and eggs. But whereas Christianity in its pure, puritanical form found it hard to assimilate these symbols, preferring instead to just impose its own festival of rebirth on top of the pagan one, capitalism, despite being in so many ways compatible with the Judeo-Christian tradition, is essentially uninterested in spiritual matters. So even though capitalism is mostly pretty okay with Christianity, which creates its own consumer-friendly occasions, it proves to be equally okay with paganism, as long as it can sell us the pagan symbols in a lucrative way.
In Christianity the idea of the life cycle is almost surreally reproduced in the (male) Trinity; God the Father, God the Son and God the Holy Spirit – defined by the Lateran Council of 1213 – 15 as “the Father who begets, the Son who is begotten, and the Holy Spirit who proceeds” – there’s no room for anything as earthly or earthy as motherhood. The Virgin Mary is essentially a token female presence, and one with her biological female attributes erased. And yet in every society that has worshipped under the Christian banner, child-bearing has historically only been done by women and child-raising has almost entirely been ‘women’s work’ too. Which makes you think that really, patriarchy is one of the great mysteries of humanity and the fact that it’s seen by many as the natural order of human society is even stranger.
But anyway; Easter. Easter is a mess, even to begin with; its name is pagan (Ēostre or Ôstara, Goddess of the spring) and its Christian traditions, even when embodied in the tragic idea of a man being murdered/sacrificed by being nailed to a cross, were never entrenched enough to suppress the celebratory, even frivolous feeling that spring traditionally brings. Okay, so Christ ascending to heaven is pretty celebratory without being frivolous; but as, in the UK at least, represented by a hot cross bun, with the cross on the top to represent the crucifix and even – to play up the morbid factor that is so central to Christianity – its spices that are supposed allude to the embalming of Christ’s dead body, it’s hardly solemn: it’s a bun.
On the other hand, birth, since the dawn of time and to the present day, is not just a simple cause for rejoicing and in that, the Christian tradition – although it tries to remove the aspects that seem most central to birth to us: women, labour (the word presumably wasn’t chosen accidentally) and procreation – probably tells us more about the seriousness and jeopardy of childbirth than the Easter bunny does.
St Margaret, “reborn” after being eaten by a dragon
Childbirth is the central and most fundamental human experience and, until the 20th century it was one of the most perilous ones, so naturally the church had to address it. And so there’s a ‘patron’ (interesting choice of word) saint of childbirth; clearly the Virgin Mary is too specialist to be identified with (and perhaps it would even be blasphemous to do so?) so instead there’s St Margaret. Not much help; firstly, St Margaret should surely be a ‘matron saint’ but that’s not a thing, and secondly, in herself she has nothing to do with birth, although she was presumably born. Instead she becomes the saint of childbirth through the symbolic act of bursting out of the dragon who ate her – a strange analogy but one that reflects the hazardous nature of childbirth in medieval times, when mortality rates were high, not just for babies but for their mothers. And what mother couldn’t relate to bursting out of a dragon? But Christianity’s real issue with the whole topic of birth has less to do with birth itself than how humans reproduce in the first place. Rabbits and hares may represent – in ancient cultures across the world, from Europe to Mexico and beyond – fecundity, but it’s an animal idea of fertility for its own sake that has nothing to do with the practical or emotional aspects of producing new human beings, or the legal, dynastic and financial ones that the Old Testament and the ancient world generally saw as the purpose of reproduction.
Jan & Hubert Van Eyck’s Eve from the Ghent Altarpiece (completed c 1432)
Pregnancy in Western art was a rarity until fairly recently and the puritanical ideas inherited by Victorian Christianity shaped art historical studies, to the point that people (until quite recently) tended to deny the evidence of their own eyes. Surely to believe that Jan and Hubert van Eyck’s hyper-realistic Eve – the mother of the human race – from Ghent Altarpiece (completed in 1432) just has the preferred medieval figure, rather than being pregnant, is perverse, isn’t it? Or that Mrs Arnolfini (Costanza Trenta) in the Arnolfini Portrait of 1434, who is touching her swollen stomach and who had died, presumably in childbirth – the year before this painting was completed, is just an example of that same fashionable shape, seems ridiculously far-fetched. (My favourite among the many theories about the Arnolfini portrait is Margaret Koster’s – which is explored in Waldemar Januszczak’s excellent short film about the painting.)
To go back to Eve; the idea of the first woman pregnant with the first child makes more sense for the 15th century, which was neither squeamish about or embarrassed by the realities of life in the same way that the 19th and early 20th century gentlemen who codified the canon of Western art history were. It’s not impossible that she is just the medieval/gothic ideal of femininity as seen in illuminated manuscripts and carvings; small shoulders, small breasts, big hips and stomach – given an unusually realistic treatment, but it’s hard to believe that even in the 15th century the first reaction of viewers – especially given the realism of the picture – wouldn’t have been to assume she was pregnant. Culture and society has changed a lot in the intervening centuries, but biology hasn’t.
For subsequent generations, the status of women and the perils of childbirth and childhood gave pregnant women and babies a strange presence in secular art. While there’s no reason to assume that people were less caring or sentimental about their partners or their children, portraits were rarely about sentiments, but status. Portraits of women, with the rare exception of Queens, were generally portraits of wives or potential wives, and pregnancy was of crucial dynastic importance. But in times when childbirth was almost as likely to end in death as life for both mother and child, it was presumably a risky thing to record; there are not very many pregnant portraits. Maybe – I should probably have investigated this before writing it – the time a portrait took from commission to completion was also a factor that made it risky? A portrait wasn’t a particularly inexpensive thing, possibly commissioning a portrait of someone who would quite likely be dead within the next nine months felt like an iffy investment, or (to be less mercenary about it) courting bad luck? In the generations that followed, female artists – such as Elizabeth Vigee-Lebrun – could celebrate motherhood in self-portraits, but for the kind of reasons mentioned above – and because of contemporary ideas of ‘decency’ – they were hardly likely to portray themselves as obviously pregnant.
Gustav Klimt – Hope 1 (1903)
As time went on and connoiseurship and ‘art history’ became a thing I don’t think it’s too much of an exaggeration to say that the arbiters of high culture in the paternalistic (at best, misogynistic at worst) society of Europe were intimidated by the female power inherent in the creation of the human race. The other side of that coin is the (slightly titillating) sense of the beauty, magic and wonder of pregnancy that the pro-female (philogynist? There must be a word) Austrian Gustav Klimt brought to art with Hope I. Beautiful though that is, Klimt’s vision isn’t really so far from the pure virgin/corrupt whore binary of medieval times, especially when you see his beautiful female figure of hope and renewal glowing against a background of death and peril. It really only when women enter the art world in greater numbers that the symbolic and magical aspects of motherhood are reconciled with the more sombre, earthly spirituality that Christianity preferred to represent in a dying man and that pregnant women can just be pregnant women.
For me, Paula Modersohn-Becker – one of my favourite painters – is the artist of pregnancy and childbirth and a painting like her Reclining Mother and Child II (1907) shows all of the human aspects that were embodied in the contorted Christian images of the Virgin Mary, crucifixion and Christ’s rebirth. In her self-portraits, the magic of Klimt without the titillating overtones, the fragility and peril of the older periods and the prosaic facts of pregnancy and what it does, good and bad, to the body, are all acknowledged. For once, it doesn’t seem ironic, only tragic, that Modersohn-Becker would be one of the many thousands of women of her era to die from complications shortly after giving birth.
Paula Modersohn-Becker – Reclining Mother & Child II (1906)Käthe Kollwitz, 1920
But once the reality had been captured, where to go from there? Anywhere, essentially; after Paula Modersohn-Becker pregnancy becomes just a subject, if a special one; art as creation representing creation. That’s a lofty way of putting it, but for the generation of German artists that followed, ‘realism’ was the whole point, some of the time at least. If Paula Modersohn-Becker represented pregnancy from the point of view of experience, capturing both its beauty and discomfort, Otto Dix the arch-realist gives us just the discomfort. His pregnant mothers are almost all exhausted working class women, heavy, swollen, weighed down by their burden. It’s a beautifully-observed point of view, and an empathetic one, but possibly a very male one too. Although Dix claimed, possibly sincerely, “I’m not that obsessed with making representations of ugliness. Everything I’ve seen is beautiful.” he nevertheless took a definite pride in shocking viewers with his art. As he also said; “All art is exorcism. I paint dreams and visions too; the dreams and visions of my time. Painting is the effort to produce order; order in yourself. There is much chaos in me, much chaos in our time.” By the time Dix painted these pictures he was a father himself, but although his paintings of his family reveal a more tender, if just as incisive, aspect to his art. When he paints these mothers-to-be, with their hard lives in the terminally unstable Weimar Republic, he paints as a pitiless observer, knowing that his work was challenging and confrontational to the generally conservative audience of his time; a time when, like ours, forces of intolerance and conservatism were closing in on the freedom embodied in art this truthful. It’s notable that, while dealing in the same harsh realities as Dix, but with a socially conscious, rather than clinical eye, the artist Käthe Kollwitz gives her women a more studiedly pitiable, though no less ‘realistic’ aura.
But the fact that Dix’s realism, though ‘objective’ was dramatically heightened is highlighted by a comparison between two paintings, one by Dix and the other by his female student Gussy Hippold-Ahnet, painted in 1931/2 and of – I think – the same model. In Dix’s painting, his most famous painting of a pregnant woman, the mother-to-be’s face is averted, hidden in darkness and it’s her almost painful roundness and heaviness that is the focus of the picture. In Hippold-Ahnet’s painting, far less dramatically, the mother sits more or less neatly, looking big but not unhappy. It’s a less dynamic and less assured piece of work – but is it any less real? In Dix’s realism, reality is generally harsh and pitiless, with no veneer of politeness or sentimentality. But although that represents a kind of underlying truth, especially about nature, people are often savage and cruel are nevertheless just as often also polite and sentimental. Gussy’s painting seems less powerful, but she is not showing us, as Dix seems to be, a faceless being representing the eternal, but rarely-remarked-on hardship involved in the joyous business of continuing the human species. Instead, sh3 shows us a woman who happens to be pregnant; both paintings are realistic, both are objective and, as with the symbolic sacrifice of Christ and the eternally recurring Easter bunny, both display different aspects of the truth.
Since the 1920s, attitudes towards pregnancy and women have fluctuated but female artists are no longer the exception within the art world and so women in art can be women in art and not women as a symbols in art. And although male artists have continued – and why not? – to paint pregnant sitters (Lucian Freud’s Pregnant Girl is a beautiful, not uncomplicated example), not surprisingly women do it better. And while I’m not sure if my favourites – Alice Neel and Paula Rego spring to mind – add anything in terms of content and meaning to Paula Modersohn-Becker’s example, what they do add is more experience, wider experience and therefore bring a truer reflection of the source and the central experience of humanity to the world. Regardless of whether or not one believes in a god, everyone believes in that creation story; which is kind of more important than an old, bearded man, a young, sacrificed man and a bird; but it doesn’t matter, there’s room in art for everything. Anyway, enjoy your chocolate eggs.
Paula Rego – The First Mass in Brazil (1993)Bonus picture: my favourite bunny in art: detail from Piero di Cosimo’s Venus, Mars & Cupid (1505)
To start with, this was mostly about books, and I think it will end that way too. But it begins with a not terribly controversial statement; hero worship is not good. And the greatest figures in the fight for human rights or human progress of one kind or another – Martin Luther King, Jr, Emmeline Pankhurst, Gandhi – without wishing to in any way diminish their achievements – would not have achieved them alone. Rosa Parks is a genuine heroine, but if she had been the only person who believed it was wrong for African-American people to be forced to give up seats for white people, the practice would still be happening. These individuals are crucial because they are catalysts for and agents of change – but the change itself happens because people – movements of people – demand it.
a bunch of lonesome and very quarrelsome heroes
This is obviously very elementary and news to nobody, but it’s still worth remembering in times like these, when people seem to be drawn to messianic figures, or to elevate people with no such pretensions to quasi-messianic status. One of the problems with messiahs is that when they don’t fulfil the hopes of their followers, their various failures or defeats (of whatever kind) take on a cataclysmic significance far beyond the usual, human kind of setback and re-evaluation. It’s only natural to feel discouraged if your political or spiritual dreams and hopes are shattered, but it’s also important to remember that the views and opinions that you were drawn to and which you agree with belong to your too. They are likely to be shared by millions of people and the fact that they are also apparently not shared by a greater number in no way invalidates them or renders them pointless.
The history of human progress is, mostly, the history of people fighting against entrenched conservative views in order to improve the lives of all people, including, incidentally, the lives of those people they are fighting against. This obviously isn’t the case in ultimately ideological revolutions like those in France or Russia, which quickly abandoned their theoretically egalitarian positions in order to remove undesirable elements altogether, or the Nazi revolution in Germany, which never pretended to be inclusive in the first place. Hopelessness, whether cynical or Kierkegaard-ishly defiant, is a natural response to depressing times, but the biggest successes of human rights movements – from the abolition of slavery to the enfranchisement of women to the end of apartheid in South Africa to the legalisation in various countries of abortion or gay marriage – have often taken place during eras which retrospectively do not seem especially enlightened; if you believe in something, there is hope.
Rome is a place, but this is mostly about people
But if change is largely driven by mass opinion and group pressure – and it demonstrably is – why is it the individual; Rameses II, Julius Caesar, Genghis Khan, Napoleon, Garibaldi, Lenin, Hitler, the Dalai Lama, Queens, Kings, political leaders – that looms so large in the way we see events historically? Anywhere from three to six million people died in the “Napoleonic Wars” – Napoleon wasn’t one of them, his armies didn’t even win them, in the end; but they are, to posterity, his wars. There i more than one answer, and one has to do with blame, but the short answer is I think because as individuals, it is individuals that we identify with. We have a sense of other peoples’ lives, we live among other people (sounds a bit Invasion of the Bodysnatchers), but we only know our own life, and we only see the world through the window of our own perceptions.
Sara Shamma self portrait
The artist Sara Shamma – who, significantly, has undertaken many humanitarian art projects, but has also done much of her most profound work in self-portraiture – said “I think understanding a human being is like understanding the whole of humanity, and the whole universe” and the more I’ve thought about that statement the more true it seems. If we truly understand any human being, it is first, foremost and perhaps only, ourselves. And, unless you are a psychopath, in which case you have my condolences, you will recognise the traits you have – perhaps every trait you have – in other people, people who may seem otherwise almost entirely different from you. When you look at the classifications humankind has made for itself – good/bad, deadly sins, cardinal virtues – these are things we know to exist because, in varying degrees, we feel them in ourselves, and therefore recognise them in others. Even that most valued human tool, objectivity, is a human tool, just as logic, which certainly seems to explain, to our understanding at least, the way the world works, is a human idea and also an ideal. Interestingly but significantly, unlike nature, mathematics or gravity, human behaviour itself routinely defies logic. When we say – to whatever extent – that we understand the universe, what I think we mean is that we understand our own conception of it. It’s easy to talk about the universe being boundless, but not limitless, or limitless, or connected to other universes as part of a multiverse (though not easy to talk about intelligently, for me), but regardless of what is ‘out there’, what we are actually talking about is all ‘in here’, in our own brain; the universe that you talk about and think about is whatever you think it is, however you perceive it. If what you believe dictates the way you live your life it might as well be, to all intents and purposes ‘the truth’. For Stephen Hawking there were black holes in space/time, and whether or not there actually are, for a creationist there really aren’t, until the day when they impinge on our lives in anything other than a theoretical way.
This is not to say that there are no actual solid facts about (for example) the nature of the universe; but nonetheless to even prove – to us personally while alive – that anything at all continues to exist after our own death is impossible. We can of see that existence goes on after other people’s deaths, but then I can say with what I believe to be complete conviction that there is no God and that human beings are just (well I wouldn’t say “just”) a kind of sentient hourglass with the added fun that you never know how much sand it holds to start with – but that doesn’t change the fact that a whole range of Gods have made and continue to make a decisive difference to the lives of other people and therefore to the world. In that way, whether or not I believe in them, they exist.
self-empowerment
But whereas the above might sound like the background for some kind of Ayn Rand-ish radical individualism, I think the opposite is true; because if any of what I have written is correct, the key part is that it applies equally to everyone. The phrase ‘we’re all in the same boat’ is being bandied about a lot lately for pandemic-related reasons, and it’s only vaguely true as regards that particular situation. We aren’t in the same boat, or even necessarily in the same kind of body exactly, but what we as human beings do all share – broadly – is the same kind of brain. We are all individuals, and If we are conscious, we are probably self-conscious. And given that we live our – as far as we can safely tell – single earthly life as an individual human being, the idea that any of us is powerless during that lifetime is nonsense. When asked to name someone who has made a difference to the world, the first person you think of should be yourself. There would be no world as you know it without you in it, and that is not a small thing; by existing, you are changing the world. Whether for better or worse, only you can say.
Having faith in other people (or even just getting along with them) makes both your and their lives better, but the belief that one particular individual outside of yourself may be the solution to the world’s (or the country’s, etc) problems is worse than feeling powerless yourself. Not only because it can reinforce that sense of powerlessness, but because it’s blatantly untrue and (I hate to use this completely devalued word, but never mind) elitist. Also, it reduces every issue, however complex, to a finite, success-or-failure one, which is rarely how the world works. The idea of the lone hero as saviour probably has about as much validity as the idea of the lone villain as the cause of whatever ills need to be cured. Hero-worship is both logical (because we see the world from the viewpoint of “I”) and also an oddly counter-intuitive ideal to have created, since in reality as we know it, the lone individual may be us, but is largely not how we live or how things work. Human beings have structured their societies, whether on the smaller level of family or tribe, to the larger ones like political parties or nations, in terms of groups of people. But I suppose it is the same humanity that makes us aware of and empathetic to the feelings of others that makes us want to reduce ideas to their black and white, bad vs good essentials and then dress those ideas up in human clothes.
childhood favourites
And so, to books! Reading fiction and watching films and TV, it’s amazing how the larger-than-life (but also simpler and therefore ironically smaller-than-life) hero/ine vs villain, protagonist vs antagonist and – most hackneyed of all (a speciality of genre fiction since such a thing existed, and the preserve or religion and mythology before that) – the ‘chosen one’ vs ‘dark lord’ narrative continues to be employed by writers and enjoyed by generations of people (myself included*), long past the age that one becomes aware of the formulaic simplification of it.
*for people of my generation, the mention of a ‘dark lord’ immediately conjures up Star Wars and Darth Vader/The Emperor, though the ‘chosen one’ theme is thankfully underplayed in the original Star Wars trilogy. George Lucas doesn’t get much credit for the prequels, but making the chosen one becomethe dark lord is an interesting twist, even if Lucifer got there first.
Whatever its origins, it seems that people do want these kinds of figures in their lives and will settle for celebrities, athletes, even politicians in lieu of the real thing. Hitler was aware of it and cast himself in the lead heroic role, ironically becoming, to posterity, the antithesis of the character he adopted; Lenin, who by any logical reading of The Communist Manifesto should have been immune to the lure of hero worship, also cast himself in the lead role, as did most of his successors to the present day; and really, to enthusiastically espouse Marxism and then approve a monumental statue of oneself displays, at best, a shocking lack of self-awareness. The Judeo-Christian god with its demand, not only to be acknowledged as the creator of everything, but also to be actually worshipped by his creations, even in his Christian, fallible, supposedly just-like-us human form, is something of a special case, but clearly these are primordial waters to be paddling in.
Still, entertainment-wise, it took a kind of epic humbling to get even to the stage we’re at now. Heroes were once demi-gods; Gilgamesh had many adventures, overcame many enemies, but when trying to conquer death found that he could not even conquer sleep. Fallible yes, but hardly someone to identify with. And Cain killed Abel, David killed Goliath, Hercules succeeded in his twelve tasks but was eventually poisoned by the blood of a hydra, Sun Wukong the Monkey King attained immortality by mistake while drunk, Beowulf was mortally wounded in his last battle against a dragon. Cúchulainn transformed into a monstrous creature and single-handedly defeated the armies of Queen Medb. King Arthur and/or the Fisher King sleep still, to be awoken when the need for them is finally great enough. These are heroes we still recognise today and would accept in the context of a blockbuster movie or doorstop-like fantasy novel, but less so in say, a soap opera or (hopefully) on Question Time. I knew some (but not all) of these stories when I was a child, but all of them would have made sense to me because, despite the differences between the settings and the societies that produced them and that which produced me, they are not really so vastly different from most of my favourite childhood stories.
Partly that’s because some of my favourite childhood stories were those same ancient stories. But even when not reading infantilised retellings of the Greek myths (I loved the Ladybird book Famous Legends Vol. 1 with its versions of Theseus and the Minotaur and Perseus and Andromeda*) it was noticeable that not all heroes were created equal. There still were heroes of the unambiguously superhuman type (in comics most obviously; like um, Superman), but in most of the books I read, the hero who conquers all through his or her (usually his) all-round superiority was rarely the lone, or sometimes not even the main protagonist. I don’t know if it’s a consequence of Christianity (or just of literacy?) but presumably at some point people decided they preferred to identify with a hero rather than to venerate them. Perhaps stories became private rather than public when people began to read for themselves, rather than listening to stories as passed down by bards or whatever? Someone will know.
*I remember being disappointed by the Clash of the Titans film version of Medusa; too monstrous, less human, somehow undermining the horror for little me
not the original set of Narnia books I had; never quite as good without Pauline Baynes’s cover art
The first real stories that I remember (this would initially be hearing rather than reading) are probably The Hobbit, The Lion, The Witch and The Wardrobe and Charlie and the Chocolate Factory – all of which have children or quasi-children as the main characters. Narnia is a special case in that there is a ‘chosen one’ – Aslan the lion – but mostly he isn’t the main focus of the narrative, Far more shadowy, there are books I was read that I never went back to and read by myself, like Pippi Longstocking and my memory of those tends to be a few images rather than an actual story. As a very little kid I know I liked The Very Hungry Caterpillar and its ilk (also, vastly less well known, The Hungry Thing by Jan Slepian and Ann Seidler in which, as I recall, some rice would be nice said a baby sucking ice). Later, I loved Tintin and Asterix and Peanuts and Garfield as well as the usual UK comics; Beano, Dandy, Oor Wullie, The Broons, Victor and Warlord etc.
The first fiction not reliant on pictures that I remember reading for myself (probably around the Beano era) would be the Narnia series (which I already knew), Richmal Crompton’s William books and, then Biggles (already by then an antique from a very different era), some Enid Blyton (I liked the less-famous Five Find-Outers best), Lloyd Alexander’s Chronicles of Prydain, and Willard Price’s Adventure series. Mostly these were all a bit old fashioned in the 80s now that I look at them, but I tended then as now to accumulate second hand books.
Lloyd Alexander’s Chronicles of Prydain; perfect marriage of author and cover art (Brian Fround and Ken Thompson)Biggles Flies Undone! Very old even when I was young, I bought this book from a jumble sale when I was 8 or 9
There was also a small group of classics that I had that must have been condensed and re-written for kids – a little brick-like paperback of Moby-Dick (Christmas present) and old hardbacks of Robinson Crusoe, Treasure Island and Kidnapped with illustrations by Broons/Oor Wullie genius Dudley D. Watkins (bought at ‘bring and buy’ sales at Primary School). Watkins’s versions of Crusoe, Long John Silver etc are still the ones I see in my head if I think of those characters. More up to date, I also had a particular fondness for Robert Westall (The Machine Gunners, The Scarecrows, The Watch House etc) and the somewhat trashy Race Against Time adventure series by JJ Fortune. This was a very 80s concoction in which a young boy from New York called Stephen, is picked up by his (this was the initial appeal) Indiana Jones-like Uncle Richard and, unbeknownst to his parents, hauled off around the world for various implausible adventures. I liked these books so much (especially the first two that I read, The Search for Mad Jack’s Crown – bought via the Chip Book Club which our school took part in – and Duel For The Samurai Sword) that I actually, for the first and last time in my life, joined a fan club. I still have the letter somewhere, warning me as a “RAT adventurer” to be prepared to be whisked away myself. Didn’t happen yet though. And then there were gamebooks (a LOT of them), which have a special place here because they fundamentally shift the focus of the narrative back to the direct hero-conquers-all themes of ancient mythology, while adding the twist that the reader themselves is that hero.
80s Hollywood blockbuster design comes to childrens’ fiction
There were also books I wouldn’t necessarily have chosen but was given at Christmas etc, books by people like Leon Garfield (adventures set in a vividly grotty evocation of 18thand early 19thcentury London), the aforementioned Moby-Dick, a comic strip version of The Mutiny on the Bounty, a Dracula annual. Also authors who I read and loved one book by, but never got around to reading more of; Anne Pilling’s Henry’s Leg, Jan Mark’s Thunder and Lightnings ( there’s a moving article about this beautifully subtle book here), Robert Leeson’s The Third Class Genie. And then there were also things we had to read at school, which mostly didn’t make a huge impression and are just evocative titles to me now – The Boy with the Bronze Axe by Kathleen Fidler and The Kelpie’s Pearls by Molly Hunter, Ian Serralliers’s The Silver Sword, Children on the Oregon Trail by Anna Rutgers van der Loeff and The Diddakoi by Rumer Godden. What did I do as a kid apart from reading?
Anyway; that’s a lot of books. And in the vast majority of them, the conclusion of the plot relies on the main character, or main character and sidekick or team to take some kind of decisive action to solve whatever problem they have. Heroism as the ancient Greeks would have understood it may largely have vanished, but even without superhuman strength or vastly superior cunning (even the fantasy novels mentioned like Lloyd Alexander’s which do still have the chosen one/dark lord idea at their heart, tend to have a fallible, doubt-filled human type of hero rather than a demigod) there is still the idea that the individual character is what matters.
it’s hard to remember a time I didn’t know these stories
And that makes sense – something like the ‘battle of five armies’ towards the end of The Hobbit is dull enough with the inclusion of characters that the reader has come to care about. A battle between armies of nameless ciphers (think the ‘Napoleonic Wars’ sans Napoleon) would be hard to get too involved in (cue image of generals with their model battlefields moving blocks of troops about, with little or no danger to themselves). Which is fair enough – being in a battle might well feel impersonal, but reading about one can’t be, if the reader is to feel any kind of drama. And maybe this is the key point – reading is – albeit at one remove – a one-on-one activity. Stephen King likens it to telepathy between the writer and reader and that is the case – they think it, we read it and it transfers from their minds to ours. And since reading is something that people seem to think children have to be made to do, often against their will, children’s authors in particular are understandably keen to engage the reader by making them identify with one character or another.
I don’t think it’s a coincidence that the most successful writers for children from CS Lewis to Enid Blyton to JK Rowling (to name just notable British ones) have tended to make children the protagonists of their books and surround their main characters with a variety of girls and boys of varying personality types. Children’s books about children are (I find) far easier to re-read as an adult than children’s books about adults are. As an adult, even JJ Fortune’s “Stephen” rings more or less true as a mostly bored tweenager of the 80s, while his Uncle Richard seems both ridiculous and vaguely creepy. “Grown up” heroes like Biggles, very vivid when encountered as a child, seem hopelessly two-dimensional and childish as an adult; what do they DO all day, when not flying planes and shooting at the enemy?
the unasked-for Christmas present that began a few years of obsessive game-playing
I mentioned gamebooks above and they – essentially single-player role playing games, often inspired by Dungeons and Dragons – deserve special mention, partly just because in the 80s, there were so many of them. There were series’ I followed and was a completist about (up to a point) – first and best being Puffin’s Fighting Fantasy (which, when I finally lost interest in them, consisted of around 30 books), there was its spin-off Steve Jackson’s Sorcery (four books), Joe Dever and Gary Chalk’s Lone Wolf (seven or eight books), Grey Star (four books), Grailquest (I think I lost interest around vol 5 or 6), then quite a few series’ that I quite liked but didn’t follow religiously – Way of the Tiger (six books), Golden Dragon (six books), Cretan Chronicles (three books) and series’ I dipped into if I happened to come across them: Choose Your Own Adventure (essentially the first gamebook series, but they mostly weren’t in the swords & sorcery genre and felt like they were aimed at a younger readership), Demonspawn (by JH Brennan, the author of Grailquest, but much, much more difficult), Falcon (time travel) and Sagard the Barbarian (four books; the selling point being that they were by D&D co-creator Gary Gygax. They were a bit clunky compared to the UK books).
Sudden memory; even before encountering my first Fighting Fantasy book, which was Steve Jackson’s Citadel of Chaos, actually the second in the series, I had bought (the Chip club again), Edward Packard’s Exploration Infinity, which was one of the Choose Your Own Adventure series, repackaged for the UK I guess, or maybe a separate book that was later absorbed into the CYOA series? Either way, there’s a particular dreamlike atmosphere that gives me a pang of complicated melancholy nostalgia when I think of the book now.
lots of books; one hero
Putting a real person – the reader – at the centre of the action ironically dispenses with the need for “character” at all, and even in books like the Lone Wolf and, Grailquest series’ where YOU are a specific person (“Lone Wolf” in the former, “Pip” in the latter), there is very little sense of (or point in) character building. You are the hero, this is what you need to do, and that’s all you need to know. In many cases, the protagonists of the heroic fantasy novels I devoured in my early teens – when I was drawn to any fat book with foil lettering and a landscape on the cover (the standard fantasy novel look in the 80s) – were not much more rounded than their lightly sketched gamebook counterparts. These books often achieved their epic length through plot only; the truly complex epic fantasy novel is a rare thing.
Thanks, presumably, to Tolkien, these plots generally revolved around main characters who were rarely ‘heroes’ in the ancient mould (though Conan and his imitators were), but were mainly inexperienced, rural quasi-children, thrust into adventures they initially had no knowledge of (Terry Brooks’s Shannara series being the classic Tolkien-lite example). But even when, as in Stephen Donaldson’s also very Tolkien-influenced Chronicles of Thomas Covenant, the hero was a cynical, unpleasant modern human being, or in Michael Moorcock’s deliberately anti-Tolkienesque Eternal Champion series, where s/he was a series of interlinked beings inhabiting the same role within different dimensions of the multiverse, the ‘chosen one’ vs some kind of implacable ‘dark lord’-ish enemy theme remains pretty constant. But this underlying core or skeleton is only most explicit in self-consciously fantastical fiction; whether or not there’s an actual dark lord or a quest, in most fiction of any kind there’s a ‘chosen one’, even if they have only been chosen by the author as the focus of the story she or he wants to tell.
Holden Caulfield and Sylvia Plath’s Esther Greenwood have this in common with Bilbo Baggins, Conan the Barbarian and William Brown. But really, what’s the alternative to books about people anyway? Even novels in which people (or surrogate people like Richard Adams’s rabbits or William Horwood’s moles) are not the main focus (or are half of the focus, like Alan Moore’s peculiar Voice of the Fire, where Northampton is essentially the ‘hero’) rely on us engaging with the writer as a writer, a human voice that becomes a kind of stand-in for a character.
classic 80s fantasy cover design
But books are not life; one of the things that unites the most undemanding pulp novelette and the greatest works of literature is that they are to some extent – like human beings – discrete, enclosed worlds; they have their beginning, middle and end. And yet, however much all of our experience relies on our perception of these key moments, that’s not necessarily how the world feels. Even complicated books are simple in that they reveal – just by seeing their length before we read them – the sense of design that is hidden from us or absent in our own lives. Even something seemingly random or illogical (the giant helmet that falls from nowhere, crushing Conrad to death in Horace Walpole’s proto-gothic novel The Castle of Otranto (1764) for example) is deliberate; recognisably something dreamlike, from the human imagination, rather than truly random as the world can be.
What we call history (“things that have happened”) usually can’t quite manage the neatness of even the most bizarre or surreal fiction. There have been genuine, almost superhuman hero/antihero/demigod figures, but how often – even when we can see their entirety – do their lives have the satisfying shape of a story? Granted, Caesar, stabbed twenty three times by his peers in the Senate chamber, has the cause-and-effect narrative of myth; but it’s an ambiguous story where the hero is the villain, depending on your point of view. Whatever one’s point of view in TheLord of the Rings or Harry Potter, to have sympathy with someone referred to (or calling themselves) a ‘dark lord’ is to consciously choose to be on the side of ‘bad’, in a way that defending a republic as a republic, or an empire as an empire isn’t.
Take Genghis Khan – ‘he’ conquered (the temptation is to also write ‘conquered’, but where do you stop with that?) – obviously not alone, but as sole leader – as much of the world as anyone has. And then, he remained successful, had issues with his succession and died in his mid 60s, in uncertain, rather than dramatic or tragic circumstances. The heroes of the Greek myths often have surprisingly downbeat endings (which I didn’t know about from the children’s versions I read) but they are usually significant in some way, and stem from the behaviour of the hero himself. Napoleon, old at 51, dying of stomach cancer or poisoning, a broken man, is not exactly a classic punishment from the Gods for hubris, or an end that anyone would have seen coming, let alone would have written for him. As ‘chosen ones’ go, Jesus is a pretty definitive example, and whether accepted as history or as fiction, he has an ending which, appropriately for god-made-man, manages to fit with both the stuff of myth (rises from the dead and ascends to heaven) but is also mundane in a way we can easily recognise; he wasn’t defeated by the Antichrist or by some supreme force of supernatural evil, but essentially killed by a committee, on the orders of someone acting against their own better judgement. More than anything else in the New Testament, that has the ring of truth to it. A significant detail too for those who want to stress the factual basis of the gospels is that the name of the murderer himself* unlike the nemeses of the ancient heroes, wasn’t even recorded.
* I guess either the guy nailing him to the cross, or the soldier spearing him in the side (much later named as Longinus, presumably for narrative purposes)
And if Jesus’s nemesis was disappointingly mundane, when on occasion, the universe does throw up something approximating a “dark lord” it doesn’t counter them with ‘chosen ones’ to defeat them either, as one might hope or expect. Living still in the shadow of WW2, Hitler’s messy and furtive end, committing suicide when beleaguered and already beaten, somehow isn’t good enough and there are a variety of rival theories about what ‘really’ happened, all of which more pleasingly fit with the kind of fiction we all grow up with. Mussolini was strung up by an angry faceless mob and his corpse was defiled. Hirohito, meanwhile, survived defeat as his troops were not supposed to do, and presided over Japan’s post-war boom to become one of the world’s longest reigning monarchs. The moral of the story is there is rarely a moral to the story. For proof of that, did the ‘heroes’ fare much better? The victors of Yalta lived on to die of a haemorrhage just months later on the eve of the unveiling of the UN (FDR), to be voted out of office, dying twenty years later a divisive figure with an ambiguous legacy (Churchill) and to become himself one of the great villains of the century with a reputation rivalling Hitler’s (Stalin).
Entertainment programs us to view history as the adventures of a series of important ‘main characters’ and how they shaped the world. It’s perhaps as good a ‘way in’ as any – like Frodo taking the ring to Mordor when no human can, or Biggles (almost) single-handedly defeating the Luftwaffe, it makes a kind of sense to us. But the distorted version of history it gives us is something to consider; think of your life and that of (name any current world leader or influential figure; apologies if you are one). If the people of the future are reading about that person, what will that tell them about your life? And what is ‘history’ telling you about really? Things that happened, yes, but prioritised by who, and for what purpose? This is an argument for reading more history, and not less I think. Other people may be the protagonists in books, but in our own personal history we have to take that role.
Artists (and historians too, in a different way) share their humanity with us, and there are great artists – you’ll have your own ideas, but William Shakespeare, Sue Townsend, Albrecht Dürer, Mickalene Thomas, Steven Spielberg and James Baldwin seems like a random but fair enough selection – who somehow have the capacity or empathy to give us insights into human beings other than (and very different from) themselves, but somehow created entirely from their own minds and their own perceptions of the world. But just like them, however aware we are of everyone else and of existence in all its variety, we can only be ourselves, and, however many boxes we seem to fit into, we can only experience the world through our own single consciousness. If there’s a chosen one, it’s you. If there’s a dark lady or a dark lord, it’s also you.
“Time, time, time, see what’s become of me…” When The Bangles covered Simon & Garfunkel’s A Hazy Shade of Winter in 1987, the song was 21 years and one month old, and now (January 2018) The Bangles’ version (from the underrated – according to me – movie of Bret Easton Ellis’ Less Than Zero) is 30 years and one month old; time flies, another year draws to an end etc etc etc. It took until the early 1990s for 60s nostalgia to really take hold and, true to form 30 years on from the 1980s, 80s nostalgia is everywhere; in music, in fashion and (especially) in film and television. Even the tired, terrifying old tropes of the cold war are back; excellent stuff.
It’s approximately 90 years since HP Lovecraft wrote, “The oldest and strongest emotion of mankind is fear, and the oldest and strongest kind of fear is the fear of the unknown.” (in the essay Supernatural Horror in Literature (1926-7)), and it’s got to be something like 25 years or so since I first read those words (in the HP Lovecraft Omnibus Vol 2, Dagon and other Macabre Tales, Grafton Books, 1985, p.423 ). So what about it?
Lovecraft might well be right about fear; but more pertinent to my intro is that possibly the oldest emotion preserved in literature – at least (major, major caveat, based on my ignorance) in the literature of Europe – is nostalgia, and the feeling that things were better in the past. (see also here for an excellent & thoughtful look at nostalgia) The literature of the ancient Greeks makes clear that the age of heroes already lay in the distant past. The pride and arrogance of Imperial Rome was tempered – formally, at least – by the belief that it was a pale imitation of the Republic which the Empire supplanted. The earliest literature in (old) English makes it clear that the inhabitants of what was one day to become England were a) not entirely sure of what had come before, but b) knew that it was in many ways ‘better’ and certainly more impressive than the present day of the 8th century:
“The work of the Giants, the stonesmiths,/ mouldereth… And the wielders and wrights?/Earthgrip holds them – gone, long gone”
The Ruin, (Translated by Michael Alexander, The Earliest English Poems, Penguin Classics (3rd edition, 1991, p. 2)
Even closer to home (for me), the earliest literature of Scotland, the Goddodin of the poet Aneirin, does something similar. The poem dates from somewhere from the 7th to 10th century and is written in the ancient British language now called Old Welsh (which it is of course, but it is also, geographically, old English and old Scots, since it seems to have been spoken in a far wider area than modern Wales). The Goddodin is a series of elegies mourning the loss of the warriors of the eponymous ancient kingdom (which spread roughly over what are now the modern Scottish regions of Lothian and Borders) in battle, and with them the heroic culture of their era.*
*a perennial theme that crops up in a very similar form in the Fortinbras subplot of Shakespeare’s Hamlet, preserved at one remove from the earliest known version of the story, Saxo Grammaticus’ elemental/mythological 13th century version from his Gesta Danorum. But even this is assumed to be derived from an earlier, lost source, probably Icelandic.
To say that nostalgia as opposed to fear may be mankind’s oldest emotion is problematic. Logically it’s difficult (chicken/egg innit), but also, for all of its obvious dominant ingredients – sadness/regret and happiness – a large component of nostalgia can be fear, and, specifically, Lovecraft’s ‘fear of the unknown’ (in this case the always unknowable future). That’s problematic for many reasons. In the examples of nostalgia noted above, the glamour (not intended to have its old, magical meaning, but actually that’s probably even more appropriate) attached to the past is partly because it can’t come again. If the people of “now” were as noble, heroic etc as the people of “then”, then somehow the past and the ancestors – a vital component of the values of most non-Christian and pre-Christian cultures – are devalued and not receiving their due reverence.
Although it seems almost incomprehensible to someone of my generation, there seems to be a similar, ‘don’t disrespect the ancestors’ unease nowadays in some circles that’s manifested in an unwillingness to condemn wholesale the expansion/existence of the British Empire. And really, it’s not very complicated – it is entirely possible to be impressed by and/or grateful for the innovations of the Victorian era – flushing toilets, railways and whatnot – while also seeing the culture and times for what they were; repressive, oppressive, misogynistic, racist, ignorant. It shouldn’t be difficult, because it’s happened before. Christianity made it easy for previous ages to condemn the pagan empires of Rome, Greece, Egypt and co (and indeed the ancient Arabic civilisations) without abandoning the inventions and innovations of those same ‘decadent’ civilisations. Indeed, even at the height of Christian belief in Europe, interest in the cultures of the pagan empires remained high, even if Christian scholars felt the need to inflict a version of their own value system onto their researches. There’s no reason that people now shouldn’t be able to do the same with the ages we have left behind, or are hopefully in the process of leaving behind. Yes, good things come from bad cultures or societies, but not because of the bad, but just because (most) human beings are extraordinary.
In 2017 there seemed to be – as I suppose there always must be – an ever-increasing number of warring nostalgias and counter-nostalgias, the latest being for the Russian Revolution in 1917 – a violent event, with vast and oppressive consequences and therefore definitely negative, but like most revolutions, born of aspirations and ideals which are hard to dismiss. In fact, Dickens’ famous opening to A Tale Of Two Cities seems uncannily prophetic, because Dickens – as he explicitly realised – could see that human nature and human actions remain fairly constant:
“It was the best of times, it was the worst of times, it was the age of wisdom, it was the age of foolishness, it was the epoch of belief, it was the epoch of incredulity, it was the season of Light, it was the season of Darkness, it was the spring of hope, it was the winter of despair, we had everything before us, we had nothing before us, we were all going direct to Heaven, we were all going direct the other way – in short, the period was so far like the present period, that some of its noisiest authorities insisted on its being received, for good or for evil, in the superlative degree of comparison only”
I think it’s probably true that it’s always the best of times, for somebody, in some respect. It’s certainly always the worst of times for other people; which sounds complacent or at least fatalistic, but only if one doesn’t try in some way to improve things. This kind of impersonal nostalgia – for ‘better’ times – is, necessarily selective. (in fact, all nostalgia is, because perception is selective – hmm, it seems like this just started copying the thing about realism I wrote recently, but bear with me) and relies to a large degree on ignorance and/or self-deception in order to be nostalgia at all.
History isn’t really a subject, history is everything; people, peoples, cultures, societies, but, necessarily “history” as it’s taught, or absorbed through popular culture, filters and simplifies. That’s important, because when people in Britain talk nostalgically about ‘Victorian values’ you can (usually) assume that they don’t intend any reference to the exploitation and subjugation of untold millions of people, child prostitution and child labour, the life expectancy of the average Victorian person etc. And, as always, history is more complex than its popular image. The Victorian era may be symbolised for British people by the building of railways or the expansion of the Empire, or by Jack the Ripper, or Queen Victoria being unamused, or by the establishment’s treatment of Oscar Wilde; but it was also the era that produced and shaped Jack the Ripper, Queen Victoria and of course, Wilde himself, as well as the whole decadent movement. Interestingly, Sigmund Freud was only two years younger than Wilde; an apparently value-free but perhaps significant observation.
This kind of complexity is what makes history more interesting than it’s sometimes given credit for. The Scottish Enlightenment was a wonderful, positive, outward-looking movement, but it coexisted in Scotland with a joyless, moralising and oppressive Calvinist culture. Time and nostalgia have a way of homogenising peoples and cultures. The popular idea of ancient Rome is probably one of conquest, grandeur and decadence, but what is the popular idea, if there is one, of ‘an ancient Roman’? Someone, probably a man, probably from Italy, in a toga or armour; quite likely an emperor, a soldier or a gladiator, rather than say, a merchant, clerk or farmer. But even within this fairly narrow image, a complex figure like the emperor Elagabalus (who was Syrian, teenage, possibly transgender) defeats the obvious school textbook perceptions of ‘Roman-ness’ (as, perhaps, it did for the Romans themselves). Even in our own time, the fact that older generations from the 60s/70s to the present could lament the passing of times when ‘men were men & women were women’ etc is – to say the least – extremely disingenuous. Presumably what they mean is a time when non-‘manly’ men could be openly discriminated against and/or abused and women could be expected to be quiet and submissive.
Similarly, throughout my life I have heard people – and not only right-wing people – talk about the economic success that Hitler brought to Germany. But you don’t have to be the chairperson of a financial think-tank to see that a programme of accelerated militarism that requires war in order to function isn’t really a viable economic model for anyone who doesn’t espouse the ideology of Nazism or at least some kind of Imperial expansionism. But people seemingly want to believe that if it wasn’t for all those pesky Nazi faults Hitler could have been a great leader. He couldn’t, though, because he was a real person, his beliefs were inseparable from everything he did and he really did exist and do the things he did and therefore wasn’t a great leader.
As I’ve said too many times already, history is complex, but nostalgia is too. It’s impossible to express in a single word other than itself, though interestingly, its original Greek meaning (‘homecoming pain’) is actually more specific than the word itself has come to be in English. Possibly because of this hard-to-express quality, most European languages tend to use variations of the word ‘nostalgia’ rather than having a word of their own with the same meaning. But despite that complexity, it tends to have a simplifying quality.
The reason for that simplifying is because, for many people, nostalgia equals safety. Political reactionaries always look to the past for examples of stability and they’ll always find them – but that stability is an illusion, caused by the fact that the past itself is stable, but only in the one sense of being unchangeable. And until the invention of the time machine it will remain unchangeable – but even so, through endless re-interpretations, re-evaluations and new points of view, the picture we have of it fluctuates almost daily. I think it’s fair to assume that (as Dickens implied) every ‘golden age’ masks a dark age but the temptation to look at the past fondly is hard to resist. It often seems otherwise, but people are, by and large fairly positive and want to look back with fondness, even if it’s a melancholy fondness. And the result of that is a softening and distancing of the darker aspects of history. A quote from the great Scottish singer/songwriter Alex Harvey strips away the soft-focus effect that the distorting lens of nostalgia puts on the past:
“Nobody ever won a war. A hundred thousand dead at Waterloo. No glory in that. Nobody needs that.” (quoted in Charles Shaar Murray’s Shots From The Hip, Penguin Books, 1991, p.71)
Numbers aside, that seems indisputably true; but evidently it isn’t, because people are entirely capable of being nostalgic about almost any negative event. ‘The Blitz Spirit’ is remembered fondly in Britain, because the few people still alive who remember it survived it, and because it happened decades ago and bombs are no longer raining down on the UK. Cinema and television is full of nostalgia for even the darkest times, largely because people are supposed to be entertained by these things and structures, stories and likeable characters are imposed on the past to make it controllable and enjoyable. And that’s just as true for the harrowing ‘war is hell’ type of film as it was for The Dirty Dozen and just as true of the revisionist ‘elegiac’ Western as for the old John Wayne kind. The revisionist Westerns tend to focus on the dying days of the ‘old West’ in grimly realistic detail, but while barely acknowledging the genocide and horror that is the real historical backdrop of the period. In a way, that’s fair enough – those stories are not about that subject – but when there are not only no (or very few) films about that subject, and it is barely even acknowledged by ‘official’ narratives of taught history, it’s a stark and telling omission.
It’s my personal feeling that nothing truly good is produced by adversity, or at least that if it is, that doesn’t offset what may have been lost. Which isn’t to deny that people are amazing, resourceful, resilient and inspiring; they are. If every golden age masks a dark age, it’s probably true too that every dark age is shot through with some elements of positivity, although I won’t scrutinise that statement too closely. Countries which were colonised by the British Empire (or indeed any empire) manage to grow and assert their independence and define their own cultures, which is good; but we can never know what or who was lost when their histories were derailed.
I love blues music (and indeed the whole phenomenon of western popular music which mostly grew from it), but again; we can never know what would have been, or what would exist now, had those creative energies not been re-directed by a couple of hundred years of slavery and exploitation. Individuals are capable of achieving almost superhuman feats of bravery and resourcefulness when facing adversity; escaping from abusers and kidnappers, rescuing people from disasters etc. But no-one in their right mind would – I hope – recommend that young people undergo these kinds of ordeals in order to fully achieve their potential.
I don’t think it’s particularly useful for individuals (although governments and institutions are a very different thing) to feel guilty about the deeds of the people of the past (or to be proud of the achievements of the past, really), but I also see no need to pretend that – for example – because India has a big railway network, the British Empire did something positive by oppressing the country’s people and culture and stealing its resources. Countries that weren’t colonised by the UK or Belgium or France or Russia have railways too. Nothing good came of the British in India. India survived anyway, just as people survive catastrophes everywhere and achieve amazing things in doing so. But you don’t celebrate an earthquake because people survive it and thrive afterwards.
Lou Reed and Rachel in 1977 (Mick Rock)
So much for impersonal nostalgia – the personal kind is in many ways very similar, if less destructive. I’ve always been a nostalgic person; both for things I don’t remember, or that were long before ‘my time’ (you name it; silent movies, the 1960s, the Weimar Republic, Hong Kong cinema of the 70s, the Northern Renaissance, the Scottish Enlightenment, 80s teen movies) and, more naturally perhaps, for things within own experience. One of the things that initially made me write this was a reference in Anthony DeCurtis’ biography Lou Reed – A Life (John Murray, 2017)* about Reed’s 70s partner/muse Rachel, a fascinating figure who seems to have vanished into history. In Googling her I discovered various sites about vanishing/vanished aspects of New York and, because old photographs are endlessly fascinating, somehow segued from that to the vanished Jewish East End of London and the vanished and vanishing everything of everywhere. But if the Jewish East London of the 1960s and the underbelly of 70s New York are irretrievable, then so is one’s own childhood, not that one wants to retrieve it, exactly.
* It’s an excellent book, but one which illustrates some of my points; Lou Reed spent most of his adult life complaining about his conservative 1950s childhood, but DeCurtis himself has a more rose-tinted view of the period, saying “In stark contrast to the identity politics of today, assimilation was the order of the day…and none of Reed’s friends, Jewish or not, recall incidents of anti-Semitism or bias” (p.14) – fair enough you would think, except that just 22 pages later he also says, ‘Richard Mishkin was a fraternity brother of Allan Hyman’s in Sigma Alpha Mu, a so-called Jewish fraternity because at the time Jews were not permitted in many other fraternities.” (p.36)
Most of the polaroids etc that make up the ever-browsable (later note; but sadly diminished) Internet K-hole appear to be American, but any child of the 80s will recognise the texture and aura of the era we grew up in. When George Orwell wrote (I think in The Lion and the Unicorn, but I might be wrong; I’ll check) – “What have you in common with the child of five whose photograph your mother keeps on the mantelpiece? Nothing, except that you happen to be the same person” he was putting his finger on one of the strange paradoxes of culture, heritage and nostalgia. The memories I have of the 1980s are made up of a distorted, child’s-eye view of events and culture which is truly mine, plus things I know now that I didn’t know then, plus other peoples’ memories, TV and films. The most potent sources of nostalgia seem to be – as the makers of shows like Stranger Things and Dark, and films like Super 8 and (too many to list) are very aware – the things you didn’t notice that you had noticed, the most ephemeral details; jingles from adverts, fonts, packaging, slang.
It’s an interesting point. The fleetingness with which you experience things has nothing to do with their power as memories. I have no idea what the first horror film I saw was, but I do know that a scene on some TV show where skinheads (or possibly a single skinhead) glued a man’s hands to the wall of a lift/elevator scared me as a child and stayed with me for a long time. Maybe that was because I used to see skinheads around on the streets (you had to watch the colour of the laces in their Doc Martens to see if they were ‘bad’ skinheads or not – though they were probably kids too, I now realise). I also know now (but didn’t then) that these were the second wave of skinheads, which is why I also saw Oi! written on various walls around the town; at the time I don’t think I ever made the connection. Again, when one thinks of the impact of very small occurrences it shows how impossible a really objective view of history is. I no longer bear any high school grudges, but without really thinking about it, there are many small and/or random sneers and insults from my youth that have stayed with me in vivid detail, along with the people and places involved. Similarly (but nicer) I will eternally feel grateful to two beautiful black girls in Camden in (I think) 1990 or 91 who made remarks to me which, even at the time were ‘not politically correct’ but which pleased me immensely; it is among the very few teenage memories that boosted rather than eroded my confidence. A tiny thing, barely even an ‘incident’, but a big deal to a painfully shy adolescent. What to make of such a minor, slightly embarrassing episode? I can still vividly remember – although it was not a rarity – my whole face burning when I blushed. People often remarked on the redness of my blushes and I remember – not even slightly nostalgically – being compared to a tomato, being told I looked like I would ‘burst’ etc at high school. And thinking about it, there’s no real conclusions to draw from that memory except that real nostalgia, unlike the nostalgia industry (“it was the 70s: Buckaroo!“, to quote Alan Partridge) is particular, not general. The Camden episode includes references to youth, gender, race etc, but it has nothing important to do with any those factors and I doubt if the two girls remembered it even days later. These are not the kinds of incidents which are worthy of a biographer’s attention; but they define my youth every bit as much as the music I listened to, the sweets I remember that no longer exist, or the clothes I wore.
My particular 80s nostalgia has less to do with “the 80s” in the sense it that it appears in TV shows and films as it does a litany of gloomy-sounding things: the urban decay of 60s and 70s council estates, indoor markets, army stores, arcades, brutalist churches that harmonised with those reinforced concrete towers that the fire brigade used for practise. This is a kind of eeriness as nostalgia; reflected in my liking for empty streets and art that represents empty streets: Algernon Newton, Maurice Utrillo, Takanori Oguiss , the photography of Masataka Nakano and taken to its extreme, Giorgio de Chirico, where the emptiness isn’t truly vacant so much as it is pregnant , reminding me always of – nostalgia again – the ruined city of Charn in CS Lewis’ The Magician’s Nephew (by far my favourite Narnia book) – which made a huge impression on me as a child. Charn may even be where my liking for such things as ‘urbex’ photography, like that of Andre Govia, and of course, The Ruin, quoted way back in the first paragraph, comes from.
The Red Tower by Giorgio de ChiricoStreet scene by Takanori Oguiss
“The passing of time and all of its crimes, is making me sad again” – sadly, one of those crimes is that when I first heard that line (from Rubber Ring by The Smiths) in 1989 or thereabouts, Morrissey seemed to be on the side of the downtrodden and marginalised, whereas now he seems to be one of that increasing number of people who pretends that the mainstream of British culture is itself somehow being marginalised; which is not only patently ridiculous but impossible – and nostalgic, of course.
There’s a whole culture industry with its own cultural shorthand that has been constructed to bolster the standardised view of any given period. Nowadays, there’s whole genre of TV shows where any decade can be summed up by some B-list cultural commentator or celebrity (often not old enough to remember what they are “remembering”) saying “He/she/it were mad, weren’t they?” about some figurehead of the era. Not so great of course, when that figurehead turns out to be Jimmy Savile or Rolf Harris, at which point even nostalgia, like history, has to be revised.
The beauty of all nostalgia is that it’s selective. The 70s that Morrissey seems to feel nostalgic about (in the true, mixed feelings sense; witness the whole of Viva Hate, which I love) wasn’t ‘better’ than nowadays, but the writer of its songs was young then; he isn’t now. There are younger people who are also nostalgic about the 70s, or the 80s, because they see the partial versions of those era(s) preserved by those who were there then, or who pretend to have been. The people who mourn the loss of the blitz spirit are really no different from me wishing I’d seen the Beatles. The people who are nostalgic for the Empire will (hopefully) never have to deal with being in charge of a mass of powerless, subject people whose resources they are stealing (or be the subject of the same), but they can enjoy the things that Empire brought to all of our lives. The ‘glory’ of Empire, like the mythical ages of Greece and Rome, and the giants that the Anglo-Saxon poet pondered over, only exists now as the faded, distorted memory of a faded, distorted memory. Like the 70s, like the 80s, like the 2010s, like yesterday, they are wonderful and terrible because they can never come again.
Happy New Year!
A note on the text above:throughout this article (and many others) I refer to ‘people’ and ‘humankind’ in what is intended to be an inclusive kind of way, referring to people of all races, genders or indeed lack of gender. I probably also refer to gender in a binary sense, partly due to context, partly no doubt through laziness. However, I do have a tendency to not use the term ‘cis’, unless absolutely necessary – for me personally, the word ‘women’ includes trans women and the word men includes trans men. I don’t intend any offence by this, but I also don’t really mind if anyone is offended. I think it’s a shame that something as basic (if not simple) as a person’s gender should be a matter of opinion, but so it seems to be. My own view is that the contents of someone’s underwear is none of my business unless they explicitly make it so.