MONTHLY BLOG 75, HUMANS AS TIME-SPECIFIC STARDUST

If citing, please kindly acknowledge copyright © Penelope J. Corfield (2017)

‘We are stardust’, sang Joni Mitchell in ‘Woodstock’ (1970) soulfully and, as it transpires, presciently.1 Poets, song-lyricists and writers of scifi are often ahead of the scientific curve (as well as sometimes barking completely up the wrong tree). A.E. Housman’s Shropshire Lad (1896) had voiced a similar starry thought:

From far, from eve and morning
And yon twelve-winded sky,
The stuff of life to knit me
Blew hither: here am I.2

Today scientists confirm that the physical materials, from which humans are composed, ultimately originate from interstellar dust, which swirls throughout the cosmos. As it does so, it transports a range of elements which are the outcomes of distant galactic explosions. These prove to be vital for life on Earth. Our genetic blueprint comes in the form of the slowly evolving human genome. And our parents, whether with or (chiefly) without medical help, are our immediate progenitors. Yet, beyond that, all human bodies contain combinations of six core elements of calcium, carbon, hydrogen, nitrogen, oxygen, and phosphorus, plus five trace elements of chlorine, magnesium, potassium, sodium, and sulphur. These materials reach earth from outer space, in the form of stardust.

2017-03 No1 wallpaper_stardust3

Illus.1 Stardust Wallpaper
Source: http://wallpaperswide.com/star_dust-wallpapers.html

It’s a stirring thought, whilst simultaneously being rather romantic. This theory (with many rival variants) is known as Panspermia and has been around in scientific communities for some time.3 It seeks to explain the origins of the organic building blocks for all living beings, linking our small segment of the universe with the much, much wider context. ‘The cosmos is within us. We are made of star-stuff’, remarked the American cosmologist Carl Sagan, adding the further striking thought that: ‘We are a way for the universe to know itself’.4 And a recent study extends the analysis to stress that humans are life-long dynamic works-in-progress, absorbing star-sourced organic nutrients from plants and using them constantly in renewing all our bodily cells.5

A summary might state that all living beings exist in Space (defining that phenomenon as stretching from our immediate physical locality into the farthest reaches of Outer Space), whilst Space is simultaneously within all living beings. As George Harrison and the Beatles sang (1967): ‘Life flows on within you and without you’.6 And since the arrow of Time is integrally linked with Space, it’s as true to remark that all living beings exist in Time, just as Time is also within all living beings. No surprises there.

Space as place is the easier phenomenon to understand, as it is visible in all its three-dimensional glory and locally tangible in the form of our immediate environments. Time, by contrast, is famously intangible and invisible. However, since the two are integrally linked, it may be argued that Space/place is actually the physical manifestation of Time/temporality. So a more accurate proposition is that all living creatures exist in Space-time, just as the combined force of Space-Time is simultaneously within all living creatures.7

What does that mean for people-as-stardust? It means that, as packages of living star-dust, we are time-specific. Individuals appear at one date and depart at another. The materials that make and continually remake us come from before our living existences and survive thereafter. But those same elements combine to form specific humans at specific points in Space-time. Incidentally, that proposition remains true even for those who believe that, via reincarnation, we live multiple lives. Each period of existence is still time-specific and sequential, so that any repeated lives are lived successively rather than concurrently.

Immediately, living in Time means that exhortations to positive self-help, which offer to teach Ten Steps to Achieving Anything You Set your Mind To, don’t really mean ‘anything’.8 Humans can’t stop Time or even alter its pace by a fraction. Hence, while people might sincerely want to be younger – or older –they can’t make themselves belong to another era which is not their own.

2017-03 No2a You Can Do Anything2017-03 No2b I want to be younger

Illus.2.1 YouTube and many other variants;
2.2 Muriel’s Good Sense of Humour, from
www.facebook.com/Muriels-Good-Sense-Of-Humour-Quotes-Pics-Sayings-166181823474070/

True, individuals sometimes manage to look younger – or older – than their actual years. They can also act younger – or older. They can try every known biological, medical, sporting and psychological method to modify their age-related physical condition. They can even try to fool the world by doctoring their birth certificates, if they think that would help. Yet people still can’t make themselves to have been born (say) one hundred years earlier than their actual birth-date – or wait to be born one hundred years later – no matter how much they might prefer an epoch change.

That proposition may seem a bit too fatalistic for go-getters in today’s aspirational culture. People can’t inhabit any era in history that they want. But, viewed the other way round, there is a reverse consolation. Each individual occupies his or her own specific slice of Space-time that can be taken by no other person. Twins, triplets or children of multiple births come very close. Yet they are still birthed in sequence and occupy neighbouring spaces. And the many other non-related people, who may be born on the same day at the same split-second, occupy their own specific segments of Space-time and have their own individual destinies. People’s life-chances may fall into broad patterns which can be analysed. Each specific trajectory, however, is individual. Thus (for example) only one of all those born in Ulm (S. Germany) on 14 March 1879 went on to become world-renowned.

Some theoretical physicists argue the opposite. They speculate whether it is possible for the same mix of physical materials to recombine at different times and in different places to form duplicates, not just of the human species, but of specific individuals, as known to history. The argument, put in non-technical terms, states that: ‘In an infinite universe, anything that can happen will happen, and happen infinitely often’.9 Of course, there is a large proviso at the start of that sentence. But, given an infinite universe, there’s scope for endless permutations and material duplications.

Hence, it is suggested that at different times and places across the cosmos: ‘There may be a trillion Albert Einsteins’.10 That would mean countless physical replicas of one baby born in Ulm on 14 March 1879. What an apparent bonus for the cosmic stock of genius.

2017-03-No3-Albert_Einstein_(Nobel)

Illus.3 Multiple Images of Albert Einstein (1879-1955) aged 42, in official 1921 portrait as winner of Nobel Prize in Physics.

However, that multiple-beings argument does not take Time seriously, as an integral part of Space-time. It would be a major surprise to identity an identical physical replica of Albert Einstein, to say the least, at some other point in the galaxy … or even down the road, in Basingstoke. Yet there’s a big difference between finding a genetically identical twin and rediscovering a historically identical personage. To find a complete historical duplicate, we would have to go to a duplicate point in Time and Space.

An Einstein physically identical twin, born in another galaxy at another time, would have a different life: he would not have learned the same languages as did the 1879 Einstein, would not have learned the same mathematics and physics, would not have worked as a junior official in the Bern Patent Office, would not have encountered Jost Winteler, the Swiss linguist who wrote on language relativity and was an intellectual mentor, and so on and so forth. In short, it was not just nature but also nurture, as welded into his own lived experience, which made Einstein into the historically pre-eminent Einstein. Equally, in another era the reception of Einstein’s physics of relativity might also be completely different. Perhaps, in another galaxy many light years away, a pronouncement from a proxy-Einstein that E = mc 2 would be greeted with yawns of boredom and the remark: why labour the obvious?

While many things can be replicated within this cosmos, unidirectional Time cannot. So, while we are stardust, we are also, as living beings, time-specific combinations of stardust. Viewed one way, that’s frustrating. Viewed another way, each one has his/her personal slice of Space-time which is shared with no-one else. Optimistic/pessimistic? Either way unavoidable. Yet, in our minds, we can and do travel to different times and places … So we must cultivate our star-dusted brainpower and imaginations instead.

1 J. Mitchell, ‘Woodstock’ (1970): ‘We are stardust./ We are golden./ And we’ve got to get ourselves/ Back to the garden’.

2 A.E. Housman, ‘From Far, from Eve and Morning’, A Shropshire Lad, Poem 32 (first pub. 1896; in Penguin edn. 1956), p. 61.

3 F. Hoyle and C. Wickramasinghe, Evolution from Space (1981).

4 C. Sagan, Cosmos (1980); and see https://www.youtube.com/watch?v=Xaj407ofjNE.

5 K. and I. Schrijver, Living with the Stars: How the Human Body Is Connected to the Life Cycles of the Earth, the Planets and the Stars (2015) and interview in http://news.nationalgeographic.com/2015/01/150128-big-bang-universe-supernova-astrophysics-health-space-ngbooktalk/.

6 G. Harrison’s song on The Beatles’ Sergeant Pepper’s Lonely Hearts Club Band (1967), subsequently covered by many other musicians.

7 Again, a minority of analysts, myself included, prefer the formulation of Time-space, rather than Space-time, in recognition of the unidirectional leading power of Time. But the conventional usage is adopted here for the purposes of discussion.

8 See e.g. M. Seetubtim, ‘Ten Steps’, http://www.huffingtonpost.com/mo-seetubtim/10-steps-to-conquering-an_b_8101046.html (2015; updated 2016).

9 P. Davies, Are We Alone? Implications of the Discovery of the Extraterrestrial Life (1995), p. 97.

10 Ibid., p. 98.

For further discussion, see

To read other discussion-points, please click here

To download Monthly Blog 75 please click here

MONTHLY BLOG 74, WHY CAN’T WE THINK ABOUT SPACE WITHOUT TIME?

If citing, please kindly acknowledge copyright © Penelope J. Corfield (2017)

Well, why not? Why can’t we think about Space without Time? It’s been tried before. A persistent, though small, minority of philosophers and physicists deny the ‘reality’ of Time.1 True, they have not yet made much headway in winning the arguments. But it’s an intriguing challenge.

Space is so manifestly here and now. Look around at people, buildings, trees, clouds, the sun, the sky, the stars … And, after all what is Time? There is no agreed definition from physicists. No simple (or even complex) formula to announce that T = whatever? Why can’t we just banish it? Think of the advantages. No Time … so no hurry to finish an essay to a temporal deadline which does not ‘really’ exist. No Time … so no need to worry about getting older as the years unfold in a temporal sequence which isn’t ‘really’ happening. In the 1980s and 1990s – a time of intellectual doubt in some Western left-leaning philosophical circles – a determined onslaught upon the concept of Time was attempted by Jacques Derrida (1930-2004). He became the high-priest of temporal rejectionism. His cause could be registered somewhere under the postmodernist banner, since postmodernist thought was very hostile to the idea of history as a subject of study. It viewed it as endlessly malleable and subjective. That attitude was close to Derrida’s attitude to temporality, although not all postmodernist thinkers endorsed Derrida’s theories.2 His brand of ultra-subjective linguistic analysis, termed ‘Deconstruction’, sounded, as dramatist Yasmina Reza jokes in Art, as though it was a tough technique straight out of an engineering manual. In fact, it allowed for an endless play of subjective meanings.

For Derrida, Time was/is a purely ‘metaphysical’ concept – and he clearly did not intend that description as a compliment. Instead, he evoked an atemporal spatiality, named khōra (borrowing a term from Plato). This timeless state, which pervades the cosmos, is supposed to act both as a receptor and as a germinator of meanings. It is an eternal Present, into which all apparent temporality is absorbed.4 Any interim thoughts or feelings about Time on the part of humans would relate purely to a subjective illusion. Its meanings would, of course, have validity for them, but not necessarily for others.

So how should we think of this all-encompassing khōra? What would Space be like without Time? When asked in 1986, Derrida boldly sketched an image of khōra as a sort of sieve-like receptacle (see Fig.1).5 It was physical and tangible. Yet it was also intended to be fluid and open. Thus the receptacle would simultaneously catch, make and filter all the meanings of the world. The following extract from an explanatory letter by Derrida by no means recounts the full complexity of Derrida’s concept but gives some of the flavour:6

I propose then […] a gilded metallic object (there is gold in the passage from [Plato’s] Timaeus on the khōra […]), to be planted obliquely in the earth. Neither vertical, nor horizontal, a extremely solid frame that would resemble at once a web, a sieve, or a grill (grid) and a stringed musical instrument (piano, harp, lyre?): strings, stringed instrument, vocal chord, etc. As a grill, grid, etc., it would have a certain relationship with the filter (a telescope, or a photographic acid bath, or a machine, which has fallen from the sky, having photographed or X-rayed – filtered – an aerial view). …
2017-02-No1b-Derrida's-sketch-and-Khora

Fig. 1 (L) Derrida’s 1986 sketch of Spatiality without Time, also (R) rendered more schematically
© Centre Canadien d’Architecture/
Canadian Centre for Architecture, Montreal.

In 1987, the cerebral American architect Peter Eisenman (1932- ), whose stark works are often described as ‘deconstructive’, launched into dialogue with Derrida. They discussed giving architectural specificity to Derrida’s khōra in a public garden in Paris.8   One cannot but admire Eisenman’s daring, given the nebulousness of the key concept. Anyway, the plan (see Fig. 2) was not realised. Perhaps there was, after all, something too metaphysical in Derrida’s own vision. Moreover, the installation, if erected, would have soon shown signs of ageing: losing its gilt, weathering, acquiring moss as well as perhaps graffiti – in other words, exhibiting the handiwork of the allegedly banished Time.2017-02-No2-Model-of-Choral-Works

Fig.2 Model of Choral Works by Peter Eisenman
© Eisenman Architects. New York

So the saga took seriously the idea of banishing Time but couldn’t do it. The very words, which Derrida enjoyed deconstructing into fragmentary components, can surely convey multiple potential messages. Yet they do so in consecutive sequences, whether spoke or written, which unfold their meanings concurrently through Time.

In fact, ever since Einstein’s conceptual break-through with his theories of Relativity, we should be thinking about Time and Space as integrally linked in one continuum. Hermann Minkowski, Einstein’s intellectual ally and former tutor, made that clear: ‘Henceforth Space by itself, and Time by itself, are doomed to fade away into mere shadows, and only a kind of union of the two will preserve an independent reality’. In practice, it’s taken the world one hundred years post-Einstein to internalise the view that propositions about Time refer to Space and vice versa. Thus had Derrida managed to abolish temporality, he would have abolished spatiality along with it. It also means that scientists should not be seeking a formula for Time alone but rather for Space-Time: S-T = whatever?

Lastly, if we do want a physical monument to either Space or Time, there’s no need for a special trip to Paris. We need only look around us. The unfolding Space-Time, in which we all live, looks exactly like the entire cosmos, or, in a detailed segment of the whole, like our local home: Planet Earth.
2017-02 No3 Earth-from-Space-Vector

Fig.3 View of Planet Earth from Space
© http://boxist.com/view-of-planet-earth-in-space/

1 For anti-Time, see J. Barbour, The End of Time: The Next Revolution in Our Understanding of the Universe (1999), esp. pp. 324-5. And the reverse in R. Healey, ‘Can Physics Coherently Deny the Reality of Time?’ in C. Callender (ed.), Time, Reality and Experience (Cambridge, 2002), pp. 293-316.

2 B. Stocker, Derrida on Deconstruction (2006); A. Weiner and S.M. Wortham (eds), Encountering Derrida: Legacies and Futures of Deconstruction (2007).

3 Line of dialogue from play by Y. Reza, Art (1994).

4 D. Wood, The Deconstruction of Time (Evanstown, Ill., 2001), pp. 260-1, 269, 270-3; J. Hodge, Derrida on Time (2007); pp. ix-x, 196-203, 205-6, 213-14.

5 R. Wilken, ‘Diagrammatology’, Electronic Book Review, 2007-05-09 (2007): http://www.electronicbookreview.com/thread/electropoetics/intermingled

6 Letter from Derrida to Peter Eisenman, 30 May 1986, as cited in N. Leach (ed.), Rethinking Architecture: A Reader in Cultural Theory (1997), pp. 342-3. See also for formal diagram based on Derrida’s sketch, G. Bennington and J. Derrida, Jacques Derrida (1993), p. 406.

7 A.E. Taylor, A Commentary of Plato’s Timaeus (Oxford, 1928).

8 J. Derrida and P. Eisenman, Chora L Works, ed. J. Kipnis and T. Leeser (New York, 19997).

9 Cited in P.J. Corfield, Time and the Shape of History (2007), p. 9.

For further discussion, see

To read other discussion-points, please click here

To download Monthly Blog 74 please click here

MONTHLY BLOG 73, WRITING INTO SILENCE ABOUT TIME

If citing, please kindly acknowledge copyright © Penelope J. Corfield (2016)

Time – great subject. Lots of it around. Universal application. Paradoxical too: time flies, time crawls … time heals, time festers … time is short, time is long … time is money, time is priceless. And the list continues: humans can do time, have time, lose time, borrow time, gain time, forget time, remember time, miss time, or beat time. What a cornucopia of possibilities. Just right for the turning of the year, with its phoenix-like imagery of interlocking death and rebirth.

2017-01-no1-rising-phoenix-tattoo
What’s more, it’s a great subject for historians. Our subject focuses upon the workings of Great Time, as evidenced in human history. (Or as evidenced in cosmic history, for those who stretch Big History to cover the entire existence of the cosmos).1   So, one way or another, Time lies at the heart of all historical studies. Without temporality, there is no chronology and no sequencing; no cause and effect; no short term and no long term.

Incidentally, the reference to temporality should properly be spatio-temporality, because, since Einstein, the integral linkage of Time and Space must be understood as a given. The usual summary of that proposition is encapsulated in the terminology of Space-Time, as coined by Minkowski in 1908. But a minority of analysts, including myself, prefer Time-Space. That formulation gives the dynamic priority to temporality, which seems right.

Writing my own study of Time and the Shape of History (2007) took me many years and was thoroughly enjoyable. Mostly I worked on my own. And, having published the book and numerous related essays, I find that I’ve basically written into silence. Not complaining. Simply an observation.

Some people say vaguely: ‘how fascinating’. Or even: ‘Wow’. But mainly they don’t say anything. Even many close colleagues, with whom I’ve worked and debated for years, never mention the book. They don’t mock or laugh or give me a critique. They simply don’t mention it. That attitude is strange to me but instructive. It’s been that way for ten years, ever since the book was published. Very few reviews. And only modest sales.

Why should that be? One general reason is that Time is one of those things that’s always around but it’s so intangible and abstract that it’s taken for granted. It’s in the aether, as it were. Why bother to say more? People do write excellent books about the history of attitudes to Time, including clocks, watches and time measurement.3   And, of course, some (not many) physicists and philosopherswrite books about the evolving study of Time and the tensions/paradoxes/mysteries within the concept. But there is relatively little literature from historians on the nature of temporality, rather than on the effects of change over Time. As a result, there were very few people willing to act as publisher’s readers, before the book was published; few willing to review; few who teach anything along these lines; and few who are interested enough to read for themselves.

Yet obviously I’ve also reflected upon the qualities of the book itself. There are two major criticisms. One is that the book’s too long. In fact, at 309 pages, it’s not exceptionally lengthy. But readers tell me that they find it so. From my point of view, the length was exactly what it took for me to work through my views. I couldn’t then have written less. Maybe now I might publish a more accessible short version, with illustrations.6

My original hope was that the brief self-contained interlink sections, appearing between each big chapter, would provide different ways into reading the whole. The book does not have to be read sequentially. The main chapters are more like the spokes of a turning wheel. So the interlinks were intended as way-stations on the journey. They play with different ideas about Time, such as time travel; time cycles; time lines; time ends; time pieces; and so forth. Furthermore, I ideally wanted these interlinks to appear on different coloured paper, to alert readers to the reading options. Alas, however, that did not prove technically possible. The compromise was to print them with a different type-face; but the visual variation is not marked enough. Perhaps I should have held out for shaded paper, or distinctive margins, for the interlinks – but anyway I didn’t.

The second criticism is related to the first. Some readers do find the book hard to read. I find that verdict difficult to understand, because it’s not written in technical language. Nor are the concepts in themselves very difficult to grasp. I think it’s because the book is densely crammed throughout with information and ideas. The effect is a ‘heavy read’. Ouch! I’m deeply sorry to have written a seemingly boring book. Particularly because to me, it’s enthralling and completely the reverse of tedious.

Having said that, it’s also good to record some cheering responses. I’ve applied my three-dimensional interpretation of Time and History to a global overview essay entitled ‘Cities in Time’. People have found that instructive rather than boring.7

And I have had one truly great compliment. An early reader told me that it made her ‘think strange thoughts’. (She meant the comment in a positive way). I was thrilled. My aim is/was to get readers to look at Time and History anew.

2017-02-melting-clock
Writing into silence is a valuable learning experience. I have not stopped communicating my ideas on my ‘home period’ in research terms, the long eighteenth century. Nor have I stopped writing concept-pieces about History, Time and the long term.8   Interest in such matters is growing. Scattered evidence comes in the form of unexpected invitations from colleagues to contribute to conferences/books. Or messages from students, raising fresh questions.

Meanwhile, I’m trying even harder to make my ideas as plain and clear as I can. And I use humour wherever possible. Interestingly, there are only few jokes about Time itself (as opposed to jokes about the effects of Time) … it’s not that sort of subject. Glad to say that I can laugh at myself instead. And, yes, I’m persevering. Time isn’t going to disappear.

1 D.G. Christian, Maps of Time: An Introduction to Big History (Berkeley, Calif., 2004).

2 P.J. Corfield, Time and the Shape of History (London, 2007).

3 E.g. L. Holford-Strevens, The History of Time: A Very Short Introduction (Oxford, 2005) and a popular overview in S. Garfield, Timekeepers: How the World became Obsessed with Time (Edinburgh, 2016).

4 See the best-selling S. Hawking, A Brief History of Time: From the Big Bang to Black Holes (London, 1988); and the more accessible J. Gribbin, The Birth of Time: How We Measured the Age of the Universe (London, 1999).

5 E.g. D. Cockburn, Other Times: Philosophical Perspectives on Past, Present and Future (Cambridge, 1997); J.T. Fraser, Time, the Familiar Stranger (Amherst, Mass., 1987).

6 For shorter accounts, see P.J. Corfield, ‘History’s Big Picture in Three Dimensions’, The Historian: Journal of the Historical Association (Winter 2007), pp. 26-30; idem, ‘History Viewed Long’ (2008), for London University’s Institute of Historical Research History: Making History Website www.history.ac.uk/makinghistory/resources/articles/long_history.html; also posted within PJC website as Pdf2; and idem, ‘Teaching History’s Big Pictures: Including Continuity as well as Change’, Teaching History: Journal of the Historical Association, 136 (Sept. 2009), pp. 53-9; also posted within PJC website as Pdf3.

7 P.J. Corfield, ‘Cities in Time’, in P. Clark (ed.), The Oxford Handbook of Cities in World History (Oxford, 2013), pp. 828-46; also posted within PJC website as CorfieldPdf29.

8 P.J. Corfield, ‘History and the Temporal Turn: Returning to Causes, Effects and Diachronic Trends’, in J-F. Dunyach (ed.), Les âges de Britannia: Repenser l’histoire des mondes Britanniques – Moyen Âge-XXIe siècle (Paris, 2015), pp. 259-73; also posted within PJC website as CorfieldPdf37; idem, ‘Time and the Historians in the Age of Relativity’, in A.C.T. Geppert and T. Kössler (eds), Obsession der Gegenwart: Zeit im 20. Jahrhundert (Göttingen, 2015), pp. 71-91; also posted within PJC website as CorfieldPdf38.

For further discussion, see

To read other discussion-points, please click here

To download Monthly Blog 73 please click here

MONTHLY BLOG 72, REMEMBERING CONRAD RUSSELL, HISTORIAN of STUART BRITAIN AND ‘LAST OF THE WHIGS’

If citing, please kindly acknowledge copyright © Penelope J. Corfield (2016)

After contributing to a panel discussion on 22 September 2016 at the Institute of Historical Research, University of London, I’ve expanded my notes as follows:

When remembering my colleague Conrad Russell (1937-2004),1 the first thing that comes to mind is his utterly distinctive presence. He was an English eccentric, in full and unselfconscious bloom. In person, Conrad was tall, latterly with something of a scholar’s stoop, and always with bright, sharp eyes. But the especially memorable thing about him was his low, grave voice (‘Conrad here’, he would intone, sepulchrally, on the phone) and his slow, very precise articulation. This stately diction, combining courtesy and erudition, gave him a tremendous impact, for those who could wait to hear him out.

He once told me that his speaking manner was something that he had consciously developed, following advice given to him in his youth by his father. In fact, given his life-long wish not to be overshadowed by his famous parent, Conrad spoke very rarely about the mathematician and public intellectual Bertrand Russell (1872-1970). Conrad, the only child of Russell’s third marriage, was brought up by his mother, who lived in isolation from the rest of the family. But the eminent father had once advised his young son to formulate each sentence fully in his mind, before giving voice to each thought.2 (Not an easy thing to do). The suggestion evidently appealed to something deep within Conrad, for he embraced the slow, stately style from his youth and maintained it throughout his lifetime.

One result was that a proportion of his students, initially at London University’s Bedford College (as it then was),3 were terrified by him, although another percentage found him brilliant and immensely stimulating. Only very few disliked him. Conrad was manifestly a kindly person. He didn’t seek to score points or consciously to attract attention as an eccentric. Yet his emphatic speaking style, laced with erudite references to English politics in the 1620s, and witticisms with punch-lines in Latin, could come as a shock to undergraduates. Especially as Conrad did not just speak ‘at’ people. He wanted replies to his questions, and hoped for laughter following his jests.

Because he thought carefully before speaking, he was also wont to preface his remarks with a little exclamation, ‘Em …’, to establish his intention of contributing to the conversation, always followed by a Pinteresque pause. That technique worked well enough in some contexts. However, when Conrad took up a prestigious academic post at Yale University (1979-84), a number of his American students protested that they could not understand him. And in a society with a cultural horror of silence, Conrad’s deliberative pauses were often filled by instant chatter from others, unintentionally ousting him from the discussion. A very English figure, he admitted ruefully that he was not psychologically at ease in the USA, much as he admired his colleagues and students at Yale. Hence his relief was no secret, when he returned to the University of London, holding successive chairs at University College London (1984-90) and King’s College (1990-2003). By this time, his lecturing powers were at their full height – lucid, precise, and argumentative, all at once.

And, of course, when in 1987 he inherited his peerage as 5th Earl Russell, following the death of his half-brother, Conrad found in the House of Lords his ideal audience. They absolutely loved him. He seemed to be a voice from a bygone era, adding gravitas to every debate in which he participated. Recently, I wondered how far Conrad was reproducing his father’s spoken style, as a scion of the intellectual aristocracy in the later nineteenth century. But a check via YouTube dispelled that thought.4 There were some similarities, in that both spoke clearly and with authority. Yet Bertrand Russell’s voice was more high-pitched and his style more insouciant than that of his youngest child.

The second unmistakable feature of Conrad’s personality and intellect was his literal-mindedness. He treated every passing comment with complete seriousness. As a result, he had no small talk. His lifeline to the social world was his wife Elizabeth (née Sanders), a former student and fellow historian whom he married in 1962. She shared Conrad’s intellectual interests but was also a fluent conversationalist. At parties, Elizabeth would appear in the heart of a crowd, wielding a cigarette and speaking vivaciously. Conrad meanwhile would stand close behind her, his head slightly inclined and nodding benignly. They were well matched, remaining devoted to one another.

2016-12-no1-conrad-and-elizabeth-russell

Fig.1 Conrad and Elizabeth Russell on the stump for Labour in Paddington South (March 1966).

My own experience of Conrad’s literal-mindedness came from an occasion when we jointly interviewed a potential candidate for an undergraduate place in the History Department at Bedford College. (That was in the 1970s, before individual interviews were replaced by generic Open Days). A flustered candidate came in late, apologising that the trains were delayed. Within moments, Conrad was engaging her in an intense discussion about the running of a nationalised rail service (as British Rail then was) and the right amounts of subsidy that it should get as a proportion of GDP. The candidate gamely rallied, and did her best. But her stricken visage silently screamed: ‘all I did was mention that the train was late’.

After a while, I asked if she’d like to talk about the historical period that she was studying for A-level. Often, interview candidates became shifty at that point. On this occasion, however, my suggestion was eagerly accepted, and the candidate discoursed at some length about the financial problems of the late Tudor monarchy. Conrad was delighted with both elements of her performance; and, as we offered her a place, commented that the young were not as uninterested in complex matters of state as they were said to be. The candidate subsequently did very well – although, alas for symmetry, she did not go on to save British Rail – but I was amused at how her apparent expertise was sparked into life purely through the intensity of Conrad’s cross-questioning.

His own interest in such topical issues was part and parcel of his life-long political commitment. At that time, he was still a member of the Labour Party, having stood (unsuccessfully) as the Labour candidate for Paddington South in 1966. But Conrad was moving across the political spectrum during the 1970s. He eventually announced his shift of allegiance to the Liberals, characteristically by writing to The Times; and later, in the Lords, he took the Liberal Democrat whip. He wanted to record his change of heart, to avoid any ambiguity; and, as a Russell, he assumed that the world would want to know.

Conrad’s literalness and love of precision were qualities that made him a paradoxical historian when interrogating written documents. On the one hand, he brought a formidable focus upon the sources, shedding prior assumptions and remaining ready to challenge old interpretations. He recast seventeenth-century political and constitutional history, as one of the intellectual leaders of what became known as ‘revisionist’ history.5 He argued that there was no evidence for an inevitable clash between crown and parliament. The breakdown in their relationship, which split the MPs into divided camps, was an outcome of chance and contingency. Those were, for him, the ruling forces of history.

On the other hand, Conrad’s super-literalism led him sometimes to overlook complexities. He did not accept that people might not mean what they said – or that they might not say what they really meant at all. If the MPs declared: ‘We fear God and honour the king’, Conrad would conclude: ‘Well, there it is. They feared God and honoured the king’. Whereas one might reply, ‘Well, perhaps they were buttering up the monarch while trying to curtail his powers? And perhaps they thought it prudent not to mention that they were prepared, if need be, to fight him – especially if they thought that was God’s will’. There are often gaps within and between both words and deeds. And long-term trends are not always expressed in people’s daily language.

In case stressing his literalism and lack of small talk makes Conrad sound unduly solemn, it’s pleasant also to record a third great quality: his good humour. He was not the sort of person who had a repertoire of rollicking jokes. And his stately demeanour meant that he was not an easy man to tease. Yet, like many people who had lonely childhoods, he enjoyed the experience of being joshed by friends, chuckling agreeably when his leg was being pulled. Common jokes among the Bedford historians were directed at Conrad’s unconventional self-catered lunches (spicy sausages with jam?) or his habit of carrying everywhere a carafe of stale, green-tinged water (soluble algae, anyone?). He was delighted, even if sometimes rather bemused, by our ribbing.

Moreover, on one celebrated occasion, Conrad turned a jest against himself into a triumph. The Head of Bedford History, Professor Mike (F.M.L.) Thompson, was at some date in the mid-1970s required to appoint a Departmental Fire & Safety Officer. It marked the start of the contemporary world of regulations for everything. Mike Thompson, with his own quixotic humour, appointed Conrad Russell to the role, amidst much laughter. Not only was he the caricature of an untidy professor, living in a chaos of books and papers, but he was, like his wife Elizabeth, an inveterate chain-smoker. In fact, there were good reasons for taking proper precautions at St John’s Lodge, the handsome Regency villa where the History Department resided, since the building lacked alternative staircases for evacuation in case of emergency. Accordingly, a fire-sling was installed in Conrad’s study, high on the top floor. Then, some months later, he instituted a rare emergency drill. At the given moment, both staff and students left the building and rushed round to the back. There we witnessed Conrad, with some athleticism,6 leap into the fire-sling. He was then winched slowly to the ground, discoursing gravely, as he descended, on his favourite topic (parliamentary politics in the 1620s) – and smoking a cigarette.

2016-12-no2-st-johns-lodge-regents-pk

Fig.2 Frontage of St John’s Lodge, the Regency villa in Regent’s Park,
where the Bedford College historians taught in the 1960s and 1970s.
Conrad Russell’s room was on the top floor, at the back.

Later, Conrad referred to his years in Bedford’s History Department with great affection. Our shared accommodation in St John’s Lodge, five minutes away from the rest of the College, created a special camaraderie. The 1970s in particular were an exciting and challenging period for him, when he was refining and changing not only his politics but also his interpretation of seventeenth-century history. The revisionists attracted much attention and controversy, especially among political historians. (Economic, demographic, social and urban historians tended to stick to their own separate agendas). Collectively, the revisionists rejected the stereotypes of both ‘Whig’7 and Marxist8 explanations of long-term change. Neither the ‘march of progress’ nor the inevitable class struggle would suffice to explain the intricacies of British history. But what was the alternative big picture? Chance and contingency played a significant role in the short-term twists and turns of events. Yet the outcomes did not just emerge completely at random. In the very long run, Parliament as an institution did become politically more powerful than the monarch, even though the powers of the crown did not disappear.

By the 1990s, the next generation of political historians were beginning to revise the revisionists in turn. There were also new challenges to the discipline as a whole from postmodernist theory. In private conversation, Conrad at times worried that the revisionists’ critique of their fellow historians might be taken (wrongly) as endorsing a sceptical view that history lacks any independent meaning or validity.

Meanwhile, new research fashions were also emerging. Political history was being eclipsed by an updated social history; gender history; ethnic history; cultural history; the history of sexuality; disability history; world history; and studies of the historical meanings of identity.

Within that changing context, Conrad began to give enhanced attention to his role in the Lords. His colleagues among the Liberal Democrats appreciated the lustre he brought to their cause. In 1999 he topped the poll by his fellow peers to remain in the House, when the number of hereditary peers was drastically cut by the process of constitutional reform. And, at his funeral, Conrad Russell was mourned, with sincere regret, as the ‘last of the Whigs’.

2016-12-no3-conrad-in-lords

Fig.3 Conrad Russell, 5th Earl Russell, speaking in the House of Lords in the early twenty-first century.

There is, however, deep irony in that accolade. In political terms, it has some truth. He was proud to come from a long line of aristocrats, of impeccable social connections and Whig/Liberal views. Listening to Conrad, one could imagine hearing the voice of his great-grandfather, Lord John Russell (1792-1878), one of the Whig architects of the 1832 Reform Act. Moreover, this important strand of aristocratic liberalism was indeed coming to an end, both sociologically and politically. On the other hand, as already noted, Conrad the historian was a scourge of both Whigs and Marxists. Somehow his view of history as lacking grand trends (say, before 1689) was hard to tally with his belief in the unfolding of parliamentary liberalism thereafter.9 At very least, the interpretative differences were challenging.

Does the ultimate contrast between Conrad Russell’s Whig/Liberal politics and his polemical anti-Whig history mean that he was a deeply troubled person? Not at all. Conrad loved his life of scholarship and politics. And he loved following arguments through to their logical outcomes, even if they left him with paradoxes. Overall, he viewed his own trajectory as centrist: as a historian, opposing the Left in the 1970s when it got too radical for him, and, as a politician, opposing the Tories in the 1980s and 1990s, when they became dogmatic free-marketeers, challenging the very concept of ‘society’.

If there is such a thing as ‘nature’s lord’ to match with ‘nature’s gentleman’, then Conrad Russell was, unselfconsciously, one among their ranks. He was grand in manner yet simple in lifestyle and chivalric towards others. One of his most endearing traits was his capacity to find a ‘trace of alpha’ in even the most unpromising student. Equally, if there is such a thing as an intellectual’s intellectual, then Conrad Russell was another exemplar, although these days a chain-smoker would not be cast in the role. He was erudite and, for some critics, too much a precisian, preoccupied with minutiae. Yet he was demonstrably ready to take on big issues.

Putting all these qualities together gives us Conrad Russell, the historian and politician who was often controversial, especially in the former role, but always sincere, always intent. One of his favourite phrases, when confronted with a new fact or idea, was: ‘It gives one furiously to think’.10 And that’s what he, courteously but firmly, always did.

Conrad Sebastian Robert Russell (1937-2004), 5th Earl Russell (1987-2004), married Elizabeth Sanders (d.2003) in 1962. Their sons, Nicholas Lyulph (d.2014) and John Francis, have in turn inherited the Russell earldom but, post Britain’s 1999 constitutional reforms, not a seat in the House of Lords.

Conrad volunteered this information, in the context of a discussion between the two of us, in the early 1970s, on the subject of parental influence upon their offspring.

Merged in 1985 to become part of Royal Holloway & Bedford New College, these days known simply as Royal Holloway, University of London (RHUL), located at Egham, Surrey.

Compare the BBC Interview Face-to-Face with Bertrand Russell (1959; reissued 2012), https://www.youtube.com/watch?v=1bZv3pSaLtY with Conrad Russell’s contribution to The Lords’ Tale, Part 18 (2009), https://www.youtube.com/watch?v=lJ_u1WM7CYA.

The intellectual excitement of that era, among revisionist circles, was well conveyed by fellow-panellist, Linda Levy Peck (George Washington University, Washington, DC).

Talking of Conrad Russell’s athleticism, some of his former students drew attention to his love of cricket. He could not only carry his bat but he also bowled parabolic googlies which rose high into the sky, spinning wildly, before dropping down vertically onto the wicket behind the flailing batsman, often taking the wicket through sheer surprise.

The term ‘Whig’, first coined in 1678/9, referred to a political stance which had considerable but never universal support throughout the eighteenth and nineteenth centuries, in support of parliamentary constraints upon the unfettered powers of monarchy, a degree of religious toleration, moderate social and political reforms, and opposition to the more pro-monarchical Tories. The ‘Whig interpretation of history’, which again was never universally supported, tended to view the unfolding of British history as the gradual and inexorable march of liberal constitutionalism, toleration, technological innovation, and socio-political reforms, together termed ‘progress’.

On which, see S. Rigby, Marxism and History: A Critical Introduction (Manchester, 1987, 1998).

This point was perceptively developed by fellow-panellist, Nicholas Tyacke (University College London).

10  Conrad showed no sign of being aware (and probably would have laughed to discover) that this phrase originated with Agatha Christie’s Hercule Poirot, in Lord Edgware Dies (1933), ch.6.

For further discussion, see

To read other discussion-points, please click here

To download Monthly Blog 72 please click here

MONTHLY BLOG 67, WHAT NEXT? INTERROGATING HISTORICAL EVIDENCE

If citing, please kindly acknowledge copyright © Penelope J. Corfield (2016)

First find your evidence and check for provenance; reliability; and typicality.But what next? Here are three golden rules; and three research steps.

Rule One: Approach every source with a keen mixture of critical excitement. That is, embrace new evidence whilst being prepared to understand all its flaws and omissions. At all times, you need to sustain a vigorous mental debate between your research theories/hypotheses/questions/arguments – and your critical interrogation of the sources.

Rule Two: Think of the obvious ways to use any given source … and the not-so-obvious. Historical evidence can be used for many purposes, including those for which it was not originally intended. Be prepared to improvise and to think of different ways of using material.

Rule Three: Play fair with the evidence. That is, don’t use it to show things that it doesn’t show. Don’t misquote or mangle. Don’t use quotations taken out of context. And, while taking note of what the sources don’t say (as well as what they do), don’t let the practice of ‘reading the silence’ turn into an exercise of castigating the past for not being the present – or of interpolating your own issues into historic sources. Playing fair with the evidence means playing fair with research methods too. Keep a constant check to ensure that you don’t, unintentionally, pre-build your answers into your research procedures.

Then, when focusing upon a specific source or group of sources, there are three steps to consider in sequence.

Step One: understand the source’s context. This step is really important and requires work. Evidence gains meaning in the context of time and place. There are many handy guides to the different types of sources and their contexts.2 But, if none are available, researchers should investigate for themselves.

Finding a sheet of paper inscribed with five words, ‘William, son of John Shakespeare’, would not get a researcher very far. But finding them in the parish book of Holy Trinity, Stratford-upon-Avon, dated 26 April 1564, provides good evidence of the baptism of the world’s most famous William Shakespeare. [The actual source contains four words in Latin, as shown in Fig.1; and a later hand marked the entry with three crosses – a rather endearing sign of research excitement but one which would rightly attract the wrath of archivists today].

2016-07 No1 1564 Baptism of W Shakespeare

Fig. 1 Baptismal record of Gulielmus filius Johannis Shakspere [sic] on 26 April 1564 in Holy Trinity Church, Stratford-upon-Avon – entry marked in R margin by three crosses in a later hand. There is no evidence for Shakespeare’s actual birthday, but a patriotic tradition dating back to the eighteenth century ascribes it to 23 April – St George’s Day.

Step Two: understand the source’s characteristic style (or ‘Register’, in the terminology of literary scholarship). This step entails identifying and allowing for the characteristic style of the source(s) under examination, including their strengths and weaknesses. Again, there are guides to the common characteristics of (say) different literary genres, such as autobiographies, diaries, letters. But again, where none exist, researchers should work it out for themselves.

At a very basic level, there are obvious differences in written texts between fiction and non-fiction. Poems, stories and songs are not intended to be taken literally. And within the ranks of non-fiction, there are many different types of writings, and levels of specificity. Private thoughts expressed casually, in (say) letters and diaries, do not necessarily constitute people’s final considered views. By contrast, signed and sealed legal documents may be taken as formal statements, even while the conventional legal language brings its own restrictions.

When Shakespeare bequeathed to his wife Anne Hathaway their ‘second best bed’, he was not comparing her to a summer’s day. He was leaving her a specific item of household furniture. It can be debated whether the legacy was a considered snub or a tender personal testimonial or a utilitarian disposal of family assets or a casual after-thought.3 But the terse legal language expresses absolutely nothing about motivation.

2016-07 No2 Shakespeare's will p3

Fig.2: Extract from p. 3 of Shakespeare’s will dated March 1616, penned by a lawyer but signed by William Shakespeare. (from original in Probate Registry, Somerset House, London). The bequest I gyve vnto my wife my second best bed with the furniture, is contained in the interpolated line of text (seventh from top), indicating that it was a late addition, made after the first draft, and hence inserted very shortly before Shakespeare’s death in April 1616.

Step 3: after assessing both context and register, it’s time to savour in full the contents of any given source or set of sources. That includes every last detail. In a written text, it’s essential to study the choice of language, as well as its content(s) and meaning(s). As my former research supervisor Jack Fisher used to say: squeeze every last drop of juice from the lemon.

Moreover, while it’s wrong to read too much cosmic meaning into every passing fragment of evidence, it’s always worth remembering that some information will turn out to be more telling than others. Keep an eye open for sources which have a significance beyond their immediate remit. (Sometimes that becomes apparent only upon later reflection, sending the researcher scurrying back to the source material for a fresh appraisal.)

The possibilities are bounded by the availability of evidence. We cannot rediscover everything about the past. So, without fresh finds, it is unlikely that researchers will ever know Shakespeare’s actual date of birth. Nonetheless, the multiplication effect of multiple sources, multiple methodologies, and endless research ideas/debates, in the context of changing perspectives over time, means that historical understanding is always expanding and always being tested. The critical assessment of bounded evidence is the launching pad for unbounded knowledge. ‘There are more things in heaven and earth, Horatio …’

1 This BLOG is the pair to BLOG/ 66 (June 2016) and is equally dedicated to all past students on the Core Course of Royal Holloway (London University)’s MA in Modern History: Power, Culture, Society.

2 For a super-exemplary analysis of sources and their context, see The Proceedings of the Old Bailey: London’s Central Criminal Court, 1674-1913 on-line, co-directed by Tim Hitchcock and Robert Shoemaker: www.oldbaileyonline.org.

3 See J. Rogers, The Second Best Bed: Shakespeare’s Will in a New Light (Westport, Conn., 1993); M.S. Hedges, The Second Best Bed: In Search of Anne Hathaway (Lewes, 2000); G. Greer, Shakespeare’s Wife (2007); and BBC report www.bbc.co.uk/programmes/articles/what-will-s-will-tells-us-about-shakespeare (2016).

For further discussion, see

To read other discussion-points, please click here

To download Monthly Blog 67 please click here

MONTHLY BLOG 65, HOW DID WOMEN FIRST MANAGE TO BREAK THE GRIP OF TRADITIONAL PATRIARCHY?

If citing, please kindly acknowledge copyright © Penelope J. Corfield (2016)

Talking of taking a long time, it took centuries for women to break the grip of traditional patriarchies. How did women manage it? In a nutshell, the historical answer was (is) that literacy was the key, education the long-term provider, and the power of persuasion by both men and women which slowly turned the key.

But let’s step back for a moment to consider why the campaign was a slow one. The answer was that it was combating profound cultural traditions. There was not one single model for the rule of men. Instead, there were countless variants of male predominance which were taken absolutely for granted. The relative subordination of women seemed to be firmly established by history, economics, family relationships, biology, theology, and state power. How to break through such a combination?

The first answer, historically, was not by attacking men. That was both bad tactics and bad ideology. It raised men’s hackles, lost support for the women’s cause, and drove a wedge between fellow-humans. Thus, while there has been (is still) much male misogyny or entrenched prejudice against women, any rival strand of female misandry or systematic hostility to men has always been much weaker as a cultural tradition. It lacks the force of affronted majesty which is still expressed in contemporary misogyny, as in anonymous comments on social media.

Certainly, for many ‘lords of creation’, who espoused traditional views, the first counter-claims on behalf of women came as a deep shock. The immediate reaction was incredulous laughter. Women who spoke out on behalf of women’s rights were caricatured as bitter, frustrated old maids. A further male response was to conjure up images of the ‘vagina dentata’ – the toothed vagina of mythology. It hinted at fear of sex and/or castration anxiety. And it certainly dashed women from any maternal pedestal: their nurturing breasts being negatived by the biting fanny.
2016-05 No1 Picasso Femme (1930)

Pablo Picasso, Femme (1930).

Accordingly, one hostile male counter-attack was to denounce feminists as no more than envious man-haters. If feminists then resisted that identification, they were pushed onto the defensive. And any denials were taken as further proof of their cunningly hidden hostility.

Historically, however, the campaigns for women’s rights were rarely presented as anti-men in intention or actuality. After all, a considerable number of men were feminists from the start, just as a certain proportion of women, as well as men, were opposed. Such complications can be seen in the suffrage campaigns in the later Victorian period. Active alongside leading suffragettes were men like George Lansbury, who in 1912 resigned as Labour MP for Bow & Bromley, to stand in a by-election on a platform of votes for women. (He lost to an opponent whose slogan was ‘No Petticoat Government’.)

Meanwhile, prominent among the opponents of the suffragettes were ladies like the educational reformer Mary Augusta Ward, who wrote novels under her married name as Mrs Humphry Ward.1 She chaired the Women’s National Anti-Suffrage League (1908-10), before it amalgamated with the Men’s National League. Yet Ward did at least consider that local government was not beyond the scope of female participation.

Such intricate cross-currents explain why the process of change was historically slow and uneven. Women in fact glided into public view, initially under the radar, through the mechanism of female literacy and then through women’s writings. In the late sixteenth century, English girls first began to take up their pens in some numbers. In well-to-do households, they learned from their brothers’ tutors or from their fathers. Protestant teachings particularly favoured the spread of basic literacy, so that true Christians could read and study the Bible, which had just been translated into the vernacular Indeed, as Eales notes, the wives and daughters of clergymen were amongst England’s first cohorts of literary ladies.2 Their achievements were not seen as revolutionary (except in the eyes of a few nervous conservatives). Education, it was believed, would make these women better wives and mothers, as well as better Christians. They were not campaigning for the vote. But they were exercising their God-given brainpower.
2016-05 No2 Eighteenth-century women's literacy

Young ladies in an eighteenth-century library, being instructed by a demure governess, under a bust of Sappho – a legendary symbol of female literary creativity.

As time elapsed, however, the diffusion of female literacy proved to be the thin end of a large wedge. Girls did indeed have brainpower – in some cases exceeding that of their brothers. Why therefore should they not have access to regular education? Given that the value of Reason was becoming ever more culturally and philosophically stressed, it seemed wise for society to utilise all its resources. That indeed was the punchiest argument later used by the feminist John Stuart Mill in his celebrated essay on The Subjection of Women (1869). Fully educating the female half of the population would have the effect, he explained, of ‘doubling the mass of mental faculties available for the higher service of humanity’. Not only society collectively but also women and men individually would gain immeasurably by accessing fresh intellectual capital.3

Practical reasoning had already become appreciated at the level of the household. Throughout the eighteenth century, more and more young women were being instructed in basic literacy skills.4 These were useful as well as polite accomplishments. One anonymous text in 1739, in the name of ‘Sophia’ [the spirit of Reason], coolly drew some logical conclusions. In an urbanising and commercialising society, work was decreasingly dependent upon brute force – and increasingly reliant upon brainpower. Hence there was/is no reason why women, with the power of Reason, should not contribute alongside men. Why should there not be female lawyers, judges, doctors, scientists, University teachers, Mayors, magistrates, politicians – or even army generals and admirals?5 After all, physical strength had long ceased to be the prime qualification for military leadership. Indeed, mere force conferred no basis for either moral or political superiority. ‘Otherwise brutes would deserve pre-eminence’.6

2016-06 No3 Woman not inferior to man titlepage
There was no inevitable chain of historical progression. But, once women took up the pen, there slowly followed successive campaigns for female education, female access to the professions, female access to the franchise, female access to boardrooms, as well as (still continuing) full female participation in government, and (on the horizon) access the highest echelons of the churches and armed forces. In the very long run, the thin wedge is working. Nonetheless, it remains wise for feminists of all stripes to argue their case with sweet reason, as there are still dark fears to allay.

1 B. Harrison, Separate Spheres: The Opposition to Women’s Suffrage in Britain (1978; 2013); J. Sutherland, Mrs Humphry Ward: Eminent Victorian, Pre-Eminent Edwardian (Oxford, 1990).

2 J. Eales, ‘Female Literacy and the Social Identity of the Clergy Family in the Seventeenth Century’, Archaeologia Cantiana, 133 (2013), pp. 67-81.

3 J.S. Mill, The Subjection of Women (1869; in Everyman edn, 1929), pp. 298-9.

4 By 1801, all women in Britain’s upper and middle classes were literate, and literacy was also spreading amongst lower-class women, especially in the growing towns.

5 Anon., Woman not Inferior to Man, by Sophia, a Person of Quality (1739), pp. 36, 38, 48.

6 Ibid., p. 51.

For further discussion, see

To read other discussion-points, please click here

To download Monthly Blog 65 please click here

MONTHLY BLOG 64, WHY IS IT TAKING SO LONG TO NORMALISE THE ROLE OF WOMEN AT THE TOP IN POLITICS?

If citing, please kindly acknowledge copyright © Penelope J. Corfield (2016)

Women can certainly make it to the top of the political greasy pole. Indira Gandhi (India), Margaret Thatcher (UK), Angela Merkel (Germany) and Aung San Suu Kyi (Myanmar) are just four among several eminent examples, in very different political systems. Nonetheless, such women are still comparatively rare. Their political pre-eminence, past or present, is not ‘normal’ and taken for granted. But … why ever not? In liberal democracies it’s now a fairly long time since women got the vote and became able to stand for office. So what factors have slowed down the transition from electoral rights to political eminence?

Put like that, one fundamental answer replies that ‘these things take time’. Getting the vote is not the same as getting power. It was/is quicker and easier to change the franchise (and that took some doing) than it was/is to change the day-to-day practice of politics, which paves the way for individual participation and advancement. Over the long term, it can be argued that democracies have become relatively ‘feminised’ in the sense that issues concerning women are now routinely part of the agenda. Yet that is not the same as female participation.

Between them, women usually carry out a range of customary tasks within the household, such as childcare, dependent-adult care, home-provisioning, and housework. All of these are time-consuming and preoccupying chores, often round-the-clock. Hence, to accommodate the political participation of women, the timetables and practices of political life need to be adjusted, as does the habitual allocation of tasks within the household – and/or the delegation of some tasks to other bodies outside the household.

A very long history lies behind the assumption of an ingrained specialisation, originating in biology and sustained by cultural expectations. Women were and often still are expected to focus upon the domestic and the familial. Men take care of the ‘outside world’. A traditional phrase pronounced that: ‘Ladies have nothing to do with politics’. Such attitudes were early challenged by feminist writers. In 1834, the feisty Maria Edgeworth denounced that attitude as ‘namby-pamby’ and ‘little missy’, in her novel entitled Helen. The heroine of the story was admonished: ‘You can hardly expect, Helen, that you, as a rational being, can go through the world as it now is, without forming any opinion on points of public importance.’1

Nonetheless, great numbers of women traditionally did and do leave such things to ‘the men’. And in Edgeworth’s novel, disappointingly, her staunchly self-denying heroine does not rise to the challenge, whether by campaigning for the female franchise or for anything as halfway radical. It was very difficult, in that era, for women to propose proto-feminist views without being satirised or pilloried; and the convention against women speaking on public platforms was almost universally upheld, by women as well as by men.2

One of the central arguments against giving the vote to women was that they were ‘virtually’ represented by their menfolk. Civic activities by husbands were taken to articulate the views of the entire household. It was specialisation in practice. Man and wife were one. She had tasks at home; he spoke for both in the wider world. On such grounds, many Victorian conservatives opposed giving the vote to women, on the grounds that such a move would give married men two votes and thus be unfair to bachelors, who were left with just one. (Spinsters, meanwhile, were supposed to be ‘virtually’ represented by their fathers).

Such views depended upon an ideal-type of ‘the household’. However, in nineteenth-century Britain (as long before), there were plenty of women who either lived alone or who lived in female-headed households. And, of course, many married couples held divergent views – or followed the lead of the wife, not the husband. (Such men were unmanly and known as ‘hen-pecked’).

Yet the ideal type was precisely that, an ideal whose power did not depend upon individual case histories. Instead, it appeared founded upon biology, tradition, culture, and, for those with committed religious views, divine will. One term for this set of attitudes is ‘patriarchy’, although this concept is not a timeless verity which stands outside history, as some argue. Yet the concept and its supporting social structures certainly had enough continuing plausibility for the idea of specialist roles and separate spheres to become convincing for many, including for those women who campaigned against the female franchise.

Eventually, however, the anti-suffragists – constituting both men and women who opposed ending the traditional male monopoly – lost the argument. At a guess, most of the female vote-deniers did in time hold their noses and exercise their new electoral rights. But such attitudes showed that ‘women’ were not and are not a single social category, speaking with one voice. They may be divided by age, wealth, social class, religion, ethnicity, politics, gender orientation, marital status, maternity or childlessness, and/or their attitudes towards feminism and to men. However, the case for women’s rights does not depend upon any need for unity or homogeneity. They participate as citizens, no more or less, sharing the range of concord and disagreements that are found within any body politic.

Today, the old arguments of the anti-suffragists are being repeated and updated by a very similar mixture of venom and ridicule on the part of the anti-feminists (both male and female), who flourish with vigour, especially on social media. They express themselves sometimes ironically, sometimes amusingly, sometimes with violent misogyny.

Detail from Thomas Rowlandson’s partly affectionate, partly brutal satire of an all-woman Bluestocking Club (1815) as an unbridled cat-fight.

The mix of backlash, uncertainty, and a confused sense of upheaval, indicates how fundamental are these long-unfolding changes in the respective roles of men and women – but also how ultimately flexible are social organisations. There are some crucial biological constants, like the fact that it is women (though not all women) who give birth to children. Yet the social arrangements for child rearing are not immutable over time. Things are not set in stone – but they don’t get transformed overnight either.

So, given that, any changes designed to help women to participate fully in political life should be matched by changes to relationships and the distribution of tasks within the household. The two ‘spheres’ are not separate, but integrally linked. If one variable is to be adapted, then the others have to adapt too. And the same applies to all areas of public and private life.
2016-04 No2 Despard and Women's Freedom League

Battersea’s Charlotte Despard, leading a march of the Women’s Freedom League (1907), campaigning for the vote with her fellow sisters, in company with a representative of the next generation, in the form of a decorously dressed young boy in the front row.

Lastly, it’s important for women to press for structural changes to help women. History needs a helping hand. And female solidarity is a good antidote to anti-feminism. At the same time, however, it’s best not to be too dogmatic about the merits of all women, simply because they are women. I’m sure that there’s no female consensus on the respective political ratings of Indira Gandhi; Margaret Thatcher; Angela Merkel; and Aung San Suu Kyi. Onwards; not always upwards; but still, yes, onwards.

1 M. Edgeworth, Helen (1834; in 1890 edn), p. 260.

2 See P.J. Corfield, ‘Women and Public Speaking’, Monthly BLOG/47 (Nov. 2014).

3 For rival perspectives, see G. Lerner, The Creation of Patriarchy (Oxford, 1986); J. Bennett, History Matters: Patriarchy and the Challenge of Feminism (Manchester 2006); and S. Goldberg, Male Dominance: The Inevitability of Patriarchy (1979).

4 B.H. Harrison, Separate Spheres: The Opposition to Women’s Suffrage in Britain (2013); A.M. Benjamin, A History of the Anti-Suffrage Movement in the United States, from 1895 to 1920: Women against Equality (Lewiston, 1991).

For further discussion, see

To read other discussion-points, please click here

To download Monthly Blog 64 please click here

MONTHLY BLOG 63, THE VALUE OF VOTING – AND WHY THE PRACTICE SHOULD NOT BE MOCKED

If citing, please kindly acknowledge copyright © Penelope J. Corfield (2016)

Many more voters than previously realised cast their votes in local and national elections in eighteenth-century England. They were thereby creating – sometimes riotously and casually, but generally decorously and seriously – a culture of constitutionalism. It amounted to an emergent proto-democracy. It was not yet a full democracy, in which all adult men and women have a vote. Yet it was a culture which, importantly, chose to decide certain key disputes by casting equal votes and by then accepting the verdict of the majority.
2016-03-Election-placard1784

Satirical sketch of election placard in Westminster (1784), showing opposition candidate Charles James Fox as a wily fox with his slogan ‘the Rights of the Commons’.
Despite the brickbats and satire, Fox won.

This procedure was much safer and more rational than deciding by fighting; much more effective in winning wider consent than deciding by bribery; and much more involving for all participants than deciding by the casting of lots. Even if not all the eligible electorate actually chose to use their vote (and there were invariably some non-participants), they always had the option.

A proto-democratic culture of constitutionalism promoted public debate about the candidates and the issues, as well as a basic respect for other points of view, which might turn out to have majority support.

In the eighteenth century, the franchise was unfair and unequal, which made it a valid target for reformers. Nonetheless, in a few large constituencies with ‘open’ electorates, the number eligible to vote, via an urban freeman or rate-payer franchise, was very great. The key examples were the cities of London, Westminster, Norwich and Bristol. They had many voters (all men in this era), from a wide range of social backgrounds: from aristocrats to artisans, shopkeepers, and even some labourers. Votes were cast publicly – which meant that votes were open to challenge if the witnessing crowd doubted the eligibility of the voter – and the results were taken as registering public opinion, in the nearest the eighteenth-century constitution offered to a serious test of the views of political ‘outsiders’.

Historically, the fact that Georgian England already had a voting tradition helps to explain how the country later made the transition into full democracy so bloodlessly. Already in the eighteenth century the rudiments of the electoral process were evolving: candidate speeches; party manifestoes; electoral slogans and placards; party colours; door-to-door election canvassing; ward organisations; celebrity endorsements; shows of public support in rival demonstrations and mass meetings; close scrutiny of the voting process; declaration of the results with, upon occasion, a formal challenge and recount; and, finally, acceptance of the outcome. (The Georgian custom of chairing the successful candidate around town was not always implemented then and is today not considered obligatory).

Not only were parliamentary and civic elections contested in these large open constituencies, but during these years the practice of constitutional voting was becoming adopted in many other, different circumstances. It pointed away from the troubled civil wars of earlier times towards a calmer, safer society. Many different non-governmental institutions used the mechanism of voting, for example to determine their own membership.

For example, in numerous private clubs and societies, potential recruits had first to be nominated by one or more existing members. Then votes were cast secretly, for or against; and any candidate who was ‘blackballed’ (negatived) was declared to have lost. This practice continues in some private clubs to this day. It was (is) a particularly severe test, since the excluded candidate might have won a large majority of all votes cast. In that case, it could be accused of being anti-democratic, allowing a small group to negative the will of the majority. The moral, in all cases, is that the rules for voting are crucial in framing how each voting system works.

Other non-governmental organisations which used some version of balloting in the eighteenth century included bank management boards and charitable institutions like London’s Foundling Hospital, established in 1739. That body used voting by its Trustees to recruit new Trustees, when existing ones died or retired. Clearly, these were socially exclusive bodies. But they were upholding the convention that each vote from a valid voter has equal value and that the will of the majority should prevail.

Importantly, too, it’s known that some middling- and lower-class groups used the mechanism of voting to resolve disputes over appointments. For example, a number of Nonconformist churches chose their ministers by such means. The Congregationalists in particular valued this procedure. Potential candidates would appear before a congregation, preach a sermon, and then submit to a vote, no doubt after further behind-the-scenes canvassing and enquiries. In these ways, many people had the experience of participating equally in a collective decision to find out what the majority (including those who were not so vociferous) really wanted.

So what follows? Firstly, the culture of voting is one to be appreciated – and used. Secondly, the constitutional rules for each system of voting really matter. They should be clearly framed to allow each system to operate fairly within its remit – and the rules, once established, should not be tampered with for partisan advantage. And lastly, electors should not be summoned frivolously to the polls. That way, disillusion and apathy develop.

Look at the low turnout for elections to the European Parliament. And there’s a good reason for that. Electors know that the institution has no real power. It is not a supreme legislative or tax-raising body; and, unlike a national Parliament, no executive government is either constituted from its ranks or is scrutinised closely by it. The current arrangements do no good for either the European Union or participatory democracy. Sham elections are destructive of a genuinely civic process, which needs to be cultivated, valued, and made real – not mocked.

1 P.J. Corfield, ‘Short Summary: Proto-Democracy’, section 1.7, in E.M. Green, P.J. Corfield and C. Harvey, Elections in Metropolitan London, 1700-1850: Vol. 1 Arguments and Evidence (Bristol, 2013), pp. 55-67; also in www.londonelectoralhistory.com, section 1.7.

2 See P.J. Corfield, ‘What’s Wrong with the Old Practice of Open Voting: Standing Up to be Counted?’ BLOG no. 53 (May 2015).

3 BLOG illustration from Rowlandson’s Procession to the Hustings after a Successful Canvass (1784) in http://www.magnoliasoft.net/ms/magnoliabox/art/547698/procession-to-the-hustings-after-a-successful-canvass-no14 (detail).

4 Over time, however, the clergy’s professional qualifications tended to be emphasised over the power of congregational election: see J.W.T. Youngs, The Congregationalists: A History (New York, 1990; 1998), pp. 69-70.

For further discussion, see

To read other discussion-points, please click here

To download Monthly Blog 63 please click here

MONTHLY BLOG 62, IS THE PAST DEAD OR ALIVE? AND THE SNARES OF SUCH BINARY QUESTIONS.

If citing, please kindly acknowledge copyright © Penelope J. Corfield (2016)

Is the past dead or alive? Posing such a binary question insists upon choice; but the options constitute a false dichotomy. Nonetheless, the death of the past is often proclaimed. This BLOG examines the arguments for and against; and highlights the snares of binary thinking.

Firstly, the past, dead or alive? The ‘death of the past’ is a common, possibly reassuring notion. If you have forgotten the History dates learned at school, then don’t worry, you are in good company. Most people have. In the USA there is a sad debate entitled: ‘Is History history?’ There is at least one book entitled The Death of the Past.1 In fact, that particular study laments that people forget far too much. Nonetheless, emphatic phrases circulate in popular culture. ‘Never look back. The past is dead and buried’. ‘The bad (or good) Old Days have gone’. Something or other is irrevocably past – rendering it ‘as dead as the proverbial dodo’, which was last reliably sighted in Mauritius in 1662.2016-02 No1 Frohawk_Dodo-1905

Illus. 1: The Dodo by F.W. Frohawk,
from L.W. Rothschild’s Extinct Birds (1907).

At the same time, however, there’s a rival strand of thought, which asserts that the past is very much alive. The most famous and often quoted claim to that effect comes from William Faulkner, writing in the American Deep South in 1951, where memories and resentments from Civil War times have far from disappeared. ‘The past is never dead’, he wrote. ‘It’s not even past’. 2

Another strong statement to that effect came from Karl Marx in 1851/2. He thundered at the unpastness of the past. Revolutionary activism was constantly hampered by old thinking and old ideas: ‘The tradition of all the dead generations weighs like a nightmare upon the brain of the living’.3

Opposition to old thinking was accordingly expressed by many later Communist leaders. The ‘new’ was good and revolutionary. Antiquity was the dangerous foe. Chairman Mao’s campaign against the ‘Four Olds’ – Old Customs, Old Culture, Old Habits, Old Ideas – was a striking example, at the time of his intended Cultural Revolution in 1966.4 Yet the fact that various traditional aspects of Chinese life still persist today indicates the difficulty of uprooting very deeply embedded social attitudes, even when using the resources of a totalitarian state.

For historians, meanwhile, it’s best to reject over-simplified choices. Many things in the past (both material and intangible) have died or come to an end. Yet far from everything has shared the same fate. Ideas, languages, cultures, religions persist through Time, incorporating changes alongside continuities; biological traits evolve over immensely long periods; the structure of the cosmos unfolds over many billennia (an emergent neologism) within a measurable framework.

Hence there’s nothing like a rigid divide between past and present. They are separated by no more than a nano-second between NOW and the immediate nano-second before NOW, so that legacies/contributions from the past infuse every moment as it is lived.

Secondly, thinking in terms of binary alternatives: Having to choose between bad/old/dead versus good/new/alive is a classic example of binary thought. It is an approach commonly cultivated by activists, for example in revolutionary or apocalyptic religious movements. Are you with the great cause or against it? Such attitudes can be psychologically powerful in binding groups together.

Binaries can also be useful when assessing the strength and weakness of an argument or a proposed course of action. As bimanual creatures, we can consider the pros and cons, using the formula ‘on the one hand’ … ‘on the other hand’. Indeed, when making a case, it’s always helpful to understand the arguments against your own. That way, when facing a fundamental critic, you are prepared. (Binary options also provide a good way to bully a witness on oath: Come on, answer, Yes or No! When the truthful reply might be ‘Somewhat’ or ‘Maybe’.)

It’s even been argued that some human societies are intrinsically binary in their deepest thought patterns. Russian culture is one that has been historically so identified.5 Hence binary switching may have helped to familiarise the population with the country’s dramatic twentieth-century lurches from Tsarism to Communism and, later, back to a different form of oligarchic Democracy. (Do today’s Russians agree; or perhaps, agree somewhat?)

Either way, there is no doubt that binary thought, like binary notation, has its uses. But studying History requires the capacity to grapple with complexity alongside simplicity. Is the past dead or alive? The answer is both and neither. It falls within the embrace of ever-stable ever-fluid Time, which lives and dies simultaneously.

J.H. Plumb, The Death of the Past (1969; reissued Harmondsworth, 1973; Basingstoke, 2003).

W. Faulkner, Requiem for a Nun (1951), Act 1, sc. 3.

K. Marx, The Eighteenth Brumaire of Louis Napoleon (1851/2), in D. McClellan (ed.), Karl Marx: Selected Writings (Oxford, 1977), p. 300.

P. Clark, The Chinese Cultural Revolution: A History (Cambridge, 2008); M. Gao, The Battle for China’s Past: Mao and the Cultural Revolution (2008).

Y.M. Lotman and B.A. Uspensky, ‘Binary Models in the Dynamics of Russian Culture’, in A.D. and A.S. Nakhimovsky (eds), The Semiotics of Russian Cultural History (Ithaca, NY., 1985), pp. 30-66.

For further discussion, see

To read other discussion-points, please click here

To download Monthly Blog 62 please click here

MONTHLY BLOG 61, DOES THE STUDY OF HISTORY ‘PROGRESS’? AND HOW DOES PLURILOGUE HELP?

If citing, please kindly acknowledge copyright © Penelope J. Corfield (2016)

Does the study of History ‘progress’? That verb is cited cautiously in inverted commas, because we are all wary of over-simplified claims for historical Progress which can be deceptive, even cruelly so. But the study of History is a highly pluralistic discipline. It’s undertaken around the world by countless specialists and generalists alike. ‘History’ does not peddle its own party line. Instead, the subject rejoices in disagreements and debates. If it does ‘progress’ towards a triumphant end-point (on a journey which never stops), then it does so through pluralistic efforts and zig-zagging routes.
2016-01 No1 city on Hill C Alcuin 2005

A dulcet vision of the City Set on a Hill (2005): the ideal outcome, always sought, always elusive.

Here I argue that the study of the human past does progress, in the sense of collectively getting better sources, methodologies, agreed practices, advice handbooks, and theoretical investigations as well as smarter popularisations, text-books, and research publications – and deeper, better overall interpretations. The route, obviously, is not a step-by-step one, with each History book being better than the one before. On the contrary. I’m talking about a very long term process, between the generations, evolving since at least the eighteenth-century advent of a secular discipline of History-writing. Things, collectively, do get better.

On the way to substantiating that assertion, it’s helpful to answer two other related questions that are often raised by doubters, viz: Why do historians keep rewriting history? Why can’t they just tell it like it was and stop arguing? Two broad answers come to mind. Firstly, the debates and argument are an essential part of the process of interrogating the past. Just as History belongs to everyone, so there is no limit to the number of historical interpretations – and a good thing too. Furthermore, in every generation, there are discoveries of new sources, or new ways of using old sources, or new technologies that encourage new methodologies – let alone new questions and new approaches from new researchers (one of the major sources of change) – and, not least, the new perspectives brought about by the unfolding of History through time. Since historical research is always focused upon a moving target, then historical writing must do likewise. It is, in other words, a triumphant component of the discipline.

Even if no new evidence on a particular topic ever emerges, changing subsequent events introduce changed perspectives. For example, should Scotland leave the United Kingdom sometime soon, then interpretations of the 1707 Act of Union will change. It will no long be regarded as a brilliant compromise settlement that gained longevity and permanence – but instead as a political expedient which had a prolonged but ultimately limited shelf-life of just over 300 years (not that long in the grand scheme of things).

Yet, if historical output is always being rewritten (and, by implication, the old stuff rejected or discarded), then how can History ‘progress’? Doesn’t that mean that each generation’s writings are only good for their own day – and, after that, as dead as the fabled dodo? But, in fact, old efforts are not all discarded. Some elements may be entirely refuted or rejected or simply forgotten. Others lie fallow and then may later be revived and re-examined. But most studies remain on hand, more or less actively, in intellectual parks (traditionally known as libraries, now redoubled in websites). There they are subjects for further circulation, consultation, debate, adaptation, modification, forgetting, retrieval and, yes, at times complete oblivion – though even a forgotten work may have influenced another which remains in circulation. The pathways to knowledge are multi-circuited.

Sites of stored learning, from libraries to websites, and interactively between them

Historians don’t start by rewriting the whole subject from scratch. Instead, they build broadly upon the work of earlier generations. For that reason, as they are engaging in a discipline that focuses upon the workings of Time, historians often begin their studies with a historiographical review, analysing the past and present state of their chosen field. Even if one individual researcher is keen to embark upon polemical warfare with an influential precursor, it is rarely the case that the polemicist rejects absolutely everything written in earlier times. The framework of dates, events, chief protagonists, is (mainly) already fixed.

In effect, there is something like a continuous dialogue between the generations – except that it’s not a true dialogue, since earlier generations can’t answer back (and can’t adapt their views in the light of criticisms). So let me invent a word. There’s a plurilogue, across time and, simultaneously, between scholars from different cultures and traditions around the world.

But here’s an annoying discovery. I am not unique in my powers of linguistic invention. I’ve just googled ‘plurilogue’ to discover that it’s already the name of a recently-established international online journal, presenting reviews of philosophy and political science. In that case, I rally to claim instead that it’s a word whose time has come. Its parallel invention is an example of plurilogue in action.

Which leads me to my last point. Historians, like the practitioners of all academic disciplines, build upon work that has gone before. Even refutations or corrections constitute a form of reconstruction. An example is the collective effort and sometimes fierce debates that it has taken, over two generations, to establish reasonably reliable figures for the extent of the state-directed murders during the Holocaust.3 Similar endeavours combine to seek accurate figures for mortality in wars, or through political purges, or as a result of epidemics – often with the useful side-effect of refuting rumours, legends and propaganda claims. In terms of knowledge, that’s progress.

All the work of previous generations provides a scaffolding, which allows for new growth, development, reconfiguration, pulling down and building up. And that assessment applies not only to the work of scholars but also to the crucial input of all who work in libraries, archives, museums, art galleries, heritage associations, and everywhere that resources are preserved and curated for the use of this and future generations. Today those who are digitising historical materials are carrying out the same essential tasks in a different medium, generating wider democratic access, as well as new challenges and endless possibilities. Certainly were Denis Diderot and the Encyclopédistes living in today’s ‘Age of Wikipedia’, they would be leading the charge to put everything onto the web – and no doubt trying to enforce greater accuracy upon wikipedia.

Access to raw data alone will not, of course, make a work of history. Historians still need to grapple with their sources, with their own ideas, and with each other – as well as with their precursors. There’s a famous maxim about ‘standing on the shoulders of giants’. Reality is not so glamorous. Historians stand on the shoulders of all who went before, giant or pygmy (reputations rise and fall retrospectively too). It’s a collective thing. Plurilogue is endlessly expanding, which makes it hard work but endlessly enthralling.

1 Most historians focus upon the human past at various points during the many generations that have existed after the advent of basic literacy. But for some purposes, the subject can focus upon the entire history of the human species, incorporating the work of biologists, anthropologists, archaeologists and the misleadingly entitled ‘prehistorians’ (who study pre-literate societies), while for yet other purposes, the human past can be integrated into the history of the Earth and, indeed, the Cosmos. See e.g. D. Christian, Maps of Time: An Introduction to Big History (Berkeley, CA, 2004).

2 http://www.plurilogue.com/2011/07/welcome-to-plurilogue.html.

3 For an introduction, see >https://en.wikipedia.org/wiki/Holocaust_victims; and long list of secondary authorities cited there.

For further discussion, see

To read other discussion-points, please click here

To download Monthly Blog 61 please click here