MONTHLY BLOG 103, WHO KNOWS THESE HISTORY GRADUATES BEFORE THE CAMERAS AND MIKES IN TODAY’S MASS MEDIA?

If citing, please kindly acknowledge copyright © Penelope J. Corfield (2019)

Image © Shutterstock 178056255

Responding to the often-asked question, What do History graduates Do? I usually reply, truthfully, that they gain employment in an immense range of occupations. But this time I’ve decided to name a popular field and to cite some high-profile cases, to give specificity to my answer. The context is the labour-intensive world of the mass media. It is no surprise to find that numerous History graduates find jobs in TV and radio. They are familiar with a big subject of universal interest – the human past – which contains something for all audiences. They are simultaneously trained to digest large amounts of disparate information and ideas, before welding them into a show of coherence. And they have specialist expertise in ‘thinking long’. That hallmark perspective buffers them against undue deference to the latest fads or fashions – and indeed buffers them against the slings and arrows of both fame and adversity.

In practice, most History graduates in the mass media start and remain behind-the-scenes. They flourish as managers, programme commissioners, and producers, generally far from the fickle bright lights of public fame. Collectively, they help to steer the evolution of a fast-changing industry, which wields great cultural clout.1

There’s no one single route into such careers, just as there’s no one ‘standard’ career pattern once there. It’s a highly competitive world. And often, in terms of personpower, a rather traditionalist one. Hence there are current efforts by UK regulators to encourage a wider diversity in terms of ethnic and gender recruiting.2 Much depends upon personal initiative, perseverance, and a willingness to start at comparatively lowly levels, generally behind the scenes. It often helps as well to have some hands-on experience – whether in student or community journalism; in film or video; or in creative applications of new social media. But already-know-it-all recruits are not as welcome as those ready and willing to learn on the job.

Generally, there’s a huge surplus of would-be recruits over the number of jobs available. It’s not uncommon for History students (and no doubt many others) to dream, rather hazily, of doing something visibly ‘big’ on TV or radio. However, front-line media jobs in the public eye are much more difficult than they might seem. They require a temperament that is at once super-alert, good-humoured, sensitive to others, and quick to respond to immediate issues – and yet is simultaneously cool under fire, not easily sidetracked, not easily hoodwinked, and implacably immune from displays of personal pique and ego-grandstanding. Not an everyday combination.

It’s also essential for media stars to have a thick skin to cope with criticism. The immediacy of TV and radio creates the illusion that individual broadcasters are personally ‘known’ to the public, who therefore feel free to commend/challenge/complain with unbuttoned intensity.

Those impressive History graduates who appear regularly before the cameras and mikes are therefore a distinctly rare breed.3 (The discussion here refers to media presenters in regular employment, not to the small number of academic stars who script and present programmes while retaining full-time academic jobs – who constitute a different sort of rare breed).

Celebrated exemplars among History graduates include the TV news journalists and media personalities Kirsty Wark (b.1955) and Laura Kuenssberg (b.1976)., who are both graduates of Edinburgh University. Both have had public accolades – Wark was elected as Fellow of the Royal Society of Edinburgh in 2017 – and both face much criticism. Kuenssberg in particular, as the BBC’s first woman political editor, is walking her way warily but effectively through the Gothic-melodrama-cum-Greek-tragedy-cum-high-farce, known as Brexit.

In a different sector of the media world, the polymathic TV and radio presenter, actor, film critic and chat-show host Jonathan Ross (b.1960) is another History graduate. He began his media career young, as a child in a TV advertisement for a breakfast cereal. (His mother, an actor, put him forward for the role). Then, having studied Modern European History at London University’s School of Slavonic & Eastern European Studies, Ross worked as a TV programme researcher behind the scenes, before eventually fronting the shows. Among his varied output, he’s written a book entitled Why Do I Say These Things? (2008). This title for his stream of reminiscences highlights the tensions involved in being a ‘media personality’. On the one hand, there’s the need to keep stoking the fires of fame; but, on the other, there’s an ever-present risk of going too far and alienating public opinion.

Similar tensions accompany the careers of two further History graduates, who are famed as sports journalists. The strain of never making a public slip must be enormous. John Inverdale (b.1957), a Southampton History graduate, and Nicky Campbell (b.1961), ditto from Aberdeen, have to cope not only with the immediacy of the sporting moment but also with the passion of the fans. After a number of years, Inverdale racked up a number of gaffes. Some were unfortunate. None fatal. Nonetheless, readers of the Daily Telegraph in August 2016 were asked rhetorically, and obviously inaccurately: ‘Why Does Everyone Hate John Inverdale?’4 That sort of over-the top response indicates the pressures of life in the public eye.

Alongside his career in media, meanwhile, Nicky Campbell used his research skills to study the story of his own adoption. His book Blue-Eyed Son (2011)5 sensitively traced his extended family roots among both Protestant and Catholic communities in Ireland. His current role as a patron of the British Association for Adoption and Fostering welds this personal experience into a public role.

The final exemplar cited here is one of the most notable pioneers among women TV broadcasters. Baroness Joan Bakewell (b.1933) has had what she describes as a ‘rackety’ career. She studied first Economics and then History at Cambridge. After that, she experienced periods of considerable TV fame followed by the complete reverse, in her ‘wilderness years’.6 Yet her media skills, her stubborn persistence, and her resistance to being publicly patronised for her good looks in the 1960s, have given Bakewell media longevity. She is not afraid of voicing her views, for example in 2008 criticising the absence of older women on British TV. In her own maturity, she can now enjoy media profiles such as that in 2019 which explains: ‘Why We Love Joan Bakewell’.7 No doubt, she takes the commendations with the same pinch of salt as she took being written off in her ‘wilderness years’.

Bakewell is also known as an author; and for her commitment to civic engagement. In 2011 she was elevated to the House of Lords as a Labour peer. And in 2014 she became President of Birkbeck College, London. In that capacity, she stresses the value – indeed the necessity – of studying History. Her public lecture on the importance of this subject urged, in timely fashion, that: ‘The spirit of enquiring, of evidence-based analysis, is demanding to be heard.’8

What do these History graduates in front of the cameras and mikes have in common? Their multifarious roles as journalists, presenters and cultural lodestars indicate that there’s no straightforward pathway to media success. These multi-skilled individuals work hard for their fame and fortunes, concealing the slog behind an outer show of relaxed affability. They’ve also learned to live with the relentless public eagerness to enquire into every aspect of their lives, from health to salaries, and then to criticise the same. Yet it may be speculated that their early immersion in the study of History has stood them in good stead. As already noted, they are trained in ‘thinking long’. And they are using that great art to ‘play things long’ in career terms as well. As already noted, multi-skilled History graduates work in a remarkable variety of fields. And, among them, some striking stars appear regularly in every household across the country, courtesy of today’s mass media.

ENDNOTES:

1 O. Bennett, A History of the Mass Media (1987); P.J. Fourtie, (ed.), Media Studies, Vol. 1: Media History, Media and Society (2nd edn., Cape Town, 2007); G. Rodman, Mass Media in a Changing World: History, Industry, Controversy (New York, 2008); .

2 See Ofcom Report on Diversity and Equal Opportunities in Television (2018): https://www.ofcom.org.uk/__data/assets/pdf_file/0021/121683/diversity-in-TV-2018-report.PDF

3 Information from diverse sources, including esp. the invaluable survey by D. Nicholls, The Employment of History Graduates: A Report for the Higher Education Authority … (2005): https://www.heacademy.ac.uk/system/files/resources/employment_of_history_students_0.pdf; and short summary by D. Nicholls, ‘Famous History Graduates’, History Today, 52/8 (2002), pp. 49-51.

4 See https://www.telegraph.co.uk/olympics/2016/08/15/why-does-everyone-hate-john-inverdale?

5 N. Campbell, Blue-Eyed Son: The Story of an Adoption (2011).

6 J. Bakewell, interviewed by S. Moss, in The Guardian, 4 April 2010: https://www.theguardian.com/lifeandstyle/2010/apr/04/joan-bakewell-harold-pinter-crumpet

7 https://www.bbc.co.uk/programmes/articles/1xZlS9nh3fxNMPm5h3DZjhs/why-we-love-joan-bakewell.

8 J. Bakewell, ‘Why History Matters: The Eric Hobsbawm Lecture’ (2014): http://joanbakewell.com/history.html.

For further discussion, see

To read other discussion-points, please click here

To download Monthly Blog 103 please click here

MONTHLY BLOG 101, ARE YOU A LUMPER OR SPLITTER? HOW WELL DO YOU KNOW YOUR OWN CAST OF MIND?

If citing, please kindly acknowledge copyright © Penelope J. Corfield (2019)
The terminology, derived from Charles Darwin,1 is hardly elegant. Yet it highlights rival polarities in the intellectual cast of mind. ‘Lumpers’ seek to assemble fragments of knowledge into one big picture, while ‘splitters’ see instead complication upon complications. An earlier permutation of that dichotomy was popularised by Isaiah Berlin. In The Hedgehog and the Fox (1953), he distinguished between brainy foxes, who know many things, and intellectual hedgehogs, who apparently know one big thing.2

Fox from © Clipart 2019; Hedgehog from © GetDrawings.com (2019)

These animalian embodiments of modes of thought are derived from a fragmentary dictum from the classical Greek poet Archilochus; and they remain more fanciful than convincing. It’s not self-evident that a hedgehog’s mentality is really so overwhelmingly single-minded.3 Nor is it clear that the reverse syndrome applies particularly to foxes, which have a reputation for craft and guile.4 To make his point with reference to human thinkers, Berlin instanced the Russian novelist Leo Tolstoy as a classic ‘hedgehog’. Really? The small and prickly hedgehog hardly seems a good proxy for a grandly sweeping thinker like Tolstoy.

Those objections to Berlin’s categories, incidentally, are good examples of hostile ‘splitting’. They quibble and contradict. Sweeping generalisations are rejected. Such objections recall a dictum in a Poul Anderson sci-fi novella, when one character states gravely that: ‘I have yet to see any problem, which, when you looked at it in the right way, did not become still more complicated’.5

Arguments between aggregators/generalisers and disaggregators/sceptics, which occur in many subjects, have been particularly high-profile among historians. The lumping/splitting dichotomy was recycled in 1975 by the American J.H. Hexter.6 He accused the Marxist Christopher Hill not only of ‘lumping’ but, even worse, of deploying historical evidence selectively, to bolster a partisan interpretation. Hill replied relatively tersely.7 He rejected the charge that he did not play fair with the sources. But he proudly accepted that, through his research, he sought to find and explain meanings in history. The polarities of lumping/splitting were plain for all to see.

Historical ‘lumpers’ argue that all analysis depends upon some degree of sorting/processing/generalising, applied to disparate information. Merely itemising date after date, or fact after fact ad infinitum, would not tell anyone anything. On those dreadful occasions when lecturers do actually proceed by listing minute details one by one (for example, going through events year by year), the audience’s frustration very quickly becomes apparent.

So ‘lumpers’ like big broad interpretations. And they tend to write big bold studies, with clear long-term trends. Karl Marx’s panoramic brief survey of world history in nine pages in The Communist Manifesto was a classic piece of ‘lumping’.8 In the twentieth century, the British Marxist historian E.P. Thompson was another ‘lumper’ who sought the big picture, although he could be a combative ‘splitter’ about the faults of others.9

‘Splitters’ conversely point out that, if there were big broad-brush interpretations that were reliably apparent, they would have been discovered and accepted by now. However, the continual debates between historians in every generation indicate that grand generalisations are continually being attacked. The progression of the subject relies upon a healthy dose of disaggregation alongside aggregation. ‘Splitters’ therefore produce accounts of rich detail, complications, diversities, propounding singular rather than universal meanings, and stressing contingency over grand trends.

Sometimes critics of historical generalisations are too angry and acerbic. They can thus appear too negative and destructive. However, one of the twentieth-century historians’ most impressive splitters was socially a witty and genial man. Intellectually, however, F.J. ‘Jack’ Fisher was widely feared for his razor-sharp and trenchant demolitions of any given historical analysis. Indeed, his super-critical cast of mind had the effect of limiting his own written output to a handful of brilliant interpretative essays rather than a ‘big book’.10 (Fisher was my research supervisor. His most caustic remark to me came after reading a draft chapter: ‘There is nothing wrong with this, other than a female desire to tell all and an Oxbridge desire to tell it chronologically.’ Ouch! Fisher was not anti-woman, although he was critical of Oxbridge where I’d taken my first degree. But he used this formulation to grab my attention – and it certainly did).

Among research historians today, the temperamental/intellectual cast of mind often inclines them to ‘splitting’, partly because there are many simplistic generalisations about history in public circulation which call out for contradiction or complication. Of course, the precise distribution around the norm remains unknown. These days, I would guestimate that the profession would divide into roughly 45% ‘lumpers’, seeking big grand overviews, and 55% ‘splitters’, stressing detail, diversity, contingency. The classification, however, does depend partly on the occasion and type of output, since single-person expositions on TV and radio encourage generalisations, while round-tables and panels thrive on disagreement where splitters can come into their own.

Moreover, there are not only personal variations, depending upon circumstance, but also major oscillations in intellectual fashions within the discipline. In the later twentieth century, for example, there was a growing, though not universal, suspicion of so-called Grand Narratives (big through-time interpretations).11 The high tide of the sceptical trend known as ‘revisionism’ challenged many old generalisations and easy assumptions. Revisionists did not constitute one single school of thought. Many did favour conservative interpretations of history, but, as remains apparent today, there was and is more than one form of conservatism. That said, revisionists were generally agreed in rejecting both left-wing Marxist conflict models of revolutionary change via class struggles and liberal Whiggish linear models of evolving Progress via spreading education, constitutional rights and so forth.12

Yet the alignments were never simple (a splitterish comment from myself). Thus J.H. Hexter was a ‘splitter’ when confronting Marxists like Hill. But he was a ‘lumper’ when propounding his own Whig view of history as a process of evolving Freedom. So Hexter’s later strictures on revisionism were as fierce as was his earlier critique of Hill.13

Ideally, most research historians probably seek to find a judicious balance between ‘lumping’/‘splitting’. There is scope both for generalisations and for qualifications. After all, there is diversity within the human experience and within the cosmos. Yet there are also common themes, deep patterns, and detectable trends.

Ultimately, however, the dichotomous choice between either ‘lumping’ or ‘splitting’ is a completely false option, when pursued to its limits. Human thought, in all the disciplines, depends upon a continuous process of building/qualifying/pulling down/rebuilding/requalifying/ and so on, endlessly. With both detailed qualifications and with generalisations. An analysis built upon And+And+And+And+And would become too airy and generalised to have realistic meaning. Just as a formulation based upon But+But+But+But+But would keep negating its own negations. So, yes. Individually, it’s worth thinking about one’s own cast of mind and intellectual inclinations. (I personally enjoy both lumping and splitting, including criticising various outworn terminologies for historical periodisation).14 Furthermore, self-knowledge allows personal scope to make auto-adjustments, if deemed desirable. And then, better still, to weld the best features of ‘lumping’ and ‘splitting’ into original thought. And+But+And+Eureka.

ENDNOTES:

1 Charles Darwin in a letter dated August 1857: ‘It is good to have hair-splitters and lumpers’: see Darwin Correspondence Letter 2130 in https://www.darwinproject.ac.uk/.

2 I. Berlin, The Hedgehog and the Fox: An Essay on Tolstoy’s View of History (1953).

3 For hedgehogs, now an endangered species, see S. Coulthard, The Hedgehog Handbook (2018). If the species were to have one big message for humans today, it would no doubt be: ‘Stop destroying our habitat and support the Hedgehog Preservation Society’.

4 M. Berman, Fox Tales and Folklore (2002).

5 From P. Anderson, Call Me Joe (1957).

6 J.H. Hexter, ‘The Burden of Proof: The Historical Method of Christopher Hill’, Times Literary Supplement, 25 Oct. 1975, repr. in J.H. Hexter, On Historians: Reappraisals of Some of the Makers of Modern History (1979), pp. 227-51.

7 For Hill’s rebuttal, see The Times Literary Supplement, 7 Nov. 1975, p. 1333.

8 K. Marx and F. Engels, The Manifesto of the Communist Party (1848), Section I: ‘Bourgeois and Proletarians’, in D. McLennan (ed.), Karl Marx: Selected Writings (Oxford, 1977), pp. 222-31.

9 Among many overviews, see e.g. C. Efstathiou, E.P. Thompson: A Twentieth-Century Romantic (2015); P.J. Corfield, E.P. Thompson, Historian: An Appreciation (1993; 2018), in PJC website http://www.penelopejcorfield.co.uk/PDF’s/CorfieldPdf45.

10 See P.J. Corfield, F.J. Fisher (1908-88) and the Dialectic of Economic History (1990; 2018), in PJC website http://www.penelopejcorfield.co.uk/PDF’s/CorfieldPdf46.

11 See esp. J-F. Lyotard, The Postmodern Condition: A Report on Knowledge (Paris, 1979; in Eng. transl. 1984), p. 7, which detected ‘an incredulity toward meta-narratives’; and further discussions in G.K. Browning, Lyotard and the End of Grand Narratives (Cardiff, 2000); and A Munslow, Narrative and History (2018). Earlier Lawrence Stone, a classic historian ‘lumper’, had detected a return to narrative styles of exposition: see L. Stone, ‘The Revival of Narrative: Reflections on a New Old History’, Past & Present, 85 (1979), pp.  3-24. But in this essay Stone was detecting a decline in social-scientific styles of History-writing – not a return to old-style Grand Narratives.

12 Revisionism is sufficiently variegated to have avoided summary within one big study. But different debates are surveyed in L. Labedz (ed.), Revisionism: Essays on the History of Marxist Ideas (1962); J.M. Maddox, Hiroshima in History: The Myths of Revisionism (1974; 2011); L. Brenner, The Iron Wall: Zionist Revisionism from Jabotinsky to Shamir (1984); E. Longley, The Living Stream: Literature and Revisionism in Ireland (Newcastle upon Tyne, 1994); and M. Haynes and J. Wolfreys (eds), History and Revolution: Refuting Revisionism (2007).

13 J.H. Hexter (1910-96) founded in 1986 the Center for the History of Freedom at Washington University, USA, where he was Professor of the History of Freedom, and launched The Making of Modern Freedom series. For his views on revisionism, see J.H. Hexter, ‘Historiographical Perspectives: The Early Stuarts and Parliaments – Old Hat and the Nouvelle Vague’, Parliamentary History, 1 (1982), pp. 181-215; and analysis in W.H. Dray, ‘J.H. Hexter, Neo-Whiggism and Early Stuart Historiography’, History & Theory, 26 (1987), pp. 133-49.

14 See e.g. P.J. Corfield, ‘Primevalism: Saluting a Renamed Prehistory’, in A. Baysal, E.L. Baysal and S. Souvatzi (eds), Time and History in Prehistory (2019), pp. 265-82; and P.J. Corfield, ‘POST-Medievalism/ Modernity/ Postmodernity?’ Rethinking History, 14 (2010), pp. 379-404; also on http://www.penelopejcorfield.co.uk/PDF’s/CorfieldPdf20.

For further discussion, see

To read other discussion-points, please click here

To download Monthly Blog 101 please click here

MONTHLY BLOG 100, CONTROLLING STREET VIOLENCE & LEARNING FROM THE DEMISE OF DUELLING

If citing, please kindly acknowledge copyright © Penelope J. Corfield (2019)

Young men carrying knives today can’t simply be equated with gentlemen duelling with rapiers in the eighteenth century. There are very many obvious differences. Nonetheless, the decline and disappearance of duelling has some relevant messages for later generations, when considering how to cope with an increase in violent street confrontations.

Both themes come under the broad rubric of controlling public expressions of male violence. By the way, such a proposition does not claim violence to be purely a masculine phenomenon. Still less does it imply that all men are prone to such behaviour. Yet it remains historically the case that weaponised acts of aggression in public and semi-public places tend to be undertaken by men – and, often, by young men at that.

Duelling developed in Europe from the sixteenth century onwards as a stylised form of combat between two aggrieved individuals.1 In terms of the technology of fighting, it was linked with the advent of the light flexible rapier, instead of the heavy old broadsword. And in terms of conflict management, the challenge to a duel took the immediate heat out of a dispute, by appointing a future date and time for the aggrieved parties to appear on the ‘field of honour’.     At the appointed hour, the meeting did not turn into an instant brawl but was increasingly codified into ritual. ‘Seconds’ accompanied the combatants, to enforce the set of evolving rules and to see fair play. They were there as friendly witnesses but also, to an extent, as referees.2 In the eighteenth century, too, surgeons were often engaged to attend, so that medical attention was available if required.

Sometimes, to be sure, there were variants in the fighting format. On one occasion in 1688 two aristocratic combatants arrived, each supported by two seconds. At a given signal, all six men launched into an uninhibited sword-fight, in which all were wounded and two of the seconds died. However, such escalations were exceptional. The seconds often began the encounter by trying to reconcile the antagonists. If successful, the would-be duellists then shook hands and declared honour to be satisfied. Hence an unknown number of angry challenges never turned into outright fighting. Would-be violence in such cases had been deflected and socially contained.

Duels certainly remained a topic of both social threat and titillating gossip. They were dramatic moments, when individual destiny appeared heightened by the danger of imminent death. Later romantic novelists and film script-writers embraced the melodrama with unwearied enthusiasm. Yet the number of real-life duels in seventeenth- and eighteenth-century Britain was tiny.

No accurate records are available, since such encounters were kept semi-clandestine. Nonetheless, contemporary legal records and newspaper reports provide some clues. Scrupulous research by the historian Robert Shoemaker has identified 236 duels in the metropolitan London area between 1660 and 1830.3 In other words, there were fewer than 1.5 duels per annum on average during these 170 years. The peak duelling decades were those of the later eighteenth century. Between 1776 and 1800, there were on average 4.5 duels per annum. Yet that total emerged from a ‘greater’ London with approximately one million inhabitants in 1801. Even taking Shoemaker’s figures as a minimum, they show that duelling was much rarer in practice than its legendary status implied.

In fact, the question might be put the other way round: why were there so many duels at all, when the practice was officially deplored? The answer has relevance to today’s discussions about knife carrying. Duelling was sustained by a degree of socio-cultural acceptance by men in elite society, who were prepared to risk the legal penalties for unlawful fighting, wounding or killing. Its continuance paid tribute to the power of custom, against the law.

By the early nineteenth century in Britain, when the practice was disappearing, it was pretty much confined to young elite men of military background. However, there were three high-profile cases when very senior Tory politicians rashly took to the field. In 1798 Prime Minister William Pitt the Younger exchanged shots with his Treasurer of the Navy. (Both missed; but Pitt retired to his bed for three weeks, overcome by stress). In 1809 George Canning, the Foreign Secretary, duelled with his fellow Cabinet member, Viscount Castlereagh, Minister for War. (Castlereagh was wounded but not fatally). Most dramatically of all, in 1829 the ‘Iron Duke’ of Wellington, then Prime Minister, confronted the Earl of Winchelsea, in a row over Catholic Emancipation. (Neither was hurt; and the Duke immediately travelled to Windsor to reassure the king that his government was not suddenly leaderless).

These ill-judged episodes were signs of the acute vehemence of political confrontations in highly pressurised times. However, critics were immediately scathing. They asked pertinently enough why the populace should obey the laws when such eminent figures were potentially breaching the peace? At very least, their rash behaviour did not encourage reverence for men in high office.

Fig.2 Equestrian statue of Duke of Wellington, located in Royal Exchange Square, Glasgow: capping the statue with a traffic cone has become a source of local amusement, despite continued disapproval from Glasgow City Council and police.

Public opinion was slowly shifting against duelling. There was no guarantee that the god of battle would give victory to the disputant who was truly in the right. Fighting empowered the bellicose over the irenic. Religious and civic authorities always opposed fighting as a means of conflict resolution. Lawyers were particularly hostile. Self-help administration of justice deprived them of the business of litigation and/or arbitration. Hence in 1822 a senior law lord defined duelling as ‘an absurd and shocking remedy for private insult’.

Other voices had long been arguing that case. In 1753 the novelist Samuel Richardson strove in Sir Charles Grandison to depict a good man who declined to fight a duel, despite being strongly provoked. True, many impatient readers found this saintly hero to be somewhat priggish. But Grandison stressed that killing or maiming a rival over a point of honour was actually the reverse of honourable.4 Bourgeois good sense was triumphing over aristocratic impetuosity, although the fictional Sir Charles had a title just to soothe any anxieties over his social respectability.

Another public declaration against duelling came from the down-to-earth American inventor and civic leader Benjamin Franklin (1706-90). In 1784 he rejected the practice as both barbaric and old-fashioned: ‘It is astonishing that the murderous practice of duelling … should continue so long in vogue’. His intervention was particularly notable, in that recourse to duelling was socially more widespread in the American colonies, with their ingrained gun culture.5 And Franklin stuck to his position, refusing to rise to sundry challenges

The force of such interventions in Britain helped to render public opinion decreasingly sympathetic to duellists. One pertinent example came from 1796. Early one morning, two Americans faced each other to duel in Hyde Park. But ten swimmers in the nearby Serpentine – some of them naked – jumped out of the water and ran to stop the fight. In this particular case, they were too late; and one contestant died. Nonetheless, witnesses testified in the ensuing murder trial that the crowd, many of middling social origins, had spontaneously intervened. Public attitudes were becoming hostile. And it was that shift, rather than major changes in law or policing, which caused the practice slowly to disappear. The last fatal duel in Scotland took place in 1826; the last in England/Wales (between two exiled Frenchmen) in 1852. When Prime Minister Peel was challenged to a political duel in the 1840s he immediately refused, on the grounds that such behaviour would be ‘childish’ as well as wrong.

Viewed in terms of Britain’s historical sociology, the decline of duelling was part of a complex process of everyday demilitarisation, in the context of the slow shift from a rural to an urbanised society. Gentlemen decreasingly carried swords for other than ceremonial purposes. Canes and umbrellas came into vogue instead. Sheridan’s play The Rivals (1775) poked fun at impetuous young gentlemen who are ready to fight for their honour. Yet they are aware that ‘a sword seen in the streets of Bath would raise as great an alarm as a mad dog’, as one character remarks. The combative Irish adventurer Sir Lucius O’Trigger is lampooned – a nice touch of auto-critique from Sheridan who came from Dublin and twice fought duels himself. And the country bumpkin Bob Acres, who is egged on to fight his rival, tellingly finds his valour ‘oozing away’ when it gets to the point.6 Audiences are invited to laugh, but sympathetically.

Interestingly, by 1775 Sheridan’s play was already behind the times in terms of the technology of fighting. By the 1760s duels had come increasingly to be fought with pistols. The last known sword duel in Britain occurred in 1785. This technological updating, supplied by industrious Birmingham gun-makers, had two paradoxical effects. On the one hand, it demonstrated that the art of duelling was quick to move with the times.

On the other hand, the advent of the pistol inadvertently saved lives. The statistics collected by Robert Shoemaker showed that unequivocally. Duels with swords, among his 236 recovered examples, resulted in deaths in 22 per cent of all cases; and woundings in another 25 per cent. By contrast, it was tricky to kill a man standing at a distance, especially with early pistols which lacked rifle sights for precise aiming. Among Shoemaker’s 236 cases, as few as 7 percent of duels with pistols resulted in death; while a further 22 percent led to woundings.

Or, the point can be put the other way round. A massive 71 percent of combatants were unharmed after an exchange of pistol shots, compared with 53 per cent of duellists who were unharmed after crossing swords. In neither case did a duel guarantee a bloodbath. But pistols were a safer bet, especially after conventions established that the combatants had to stand at a considerable distance from one another and had to wait for a signal, in the form of a dropping handkerchief, before taking aim and firing. No ‘jumping the gun’. Indeed one test case in 1750 saw a duellist on trial for murder because he had fired before his opponent was ready. So the victim had testified, plaintively, on his deathbed.

It was the unavoidable proximity of the combatants rather than their martial skills which led to the greater proportion of killings by swordsmen than by gunsmen. That fact is relevant to the experience of knife-carrying today. The number of fatalities is not a sign of a special outcrop of wickedness but rather the consequence of the chosen technology. Knife-wielding in anger at close quarters is intrinsically dangerous, whatever the level of fighting expertise.

Needless to say, the moral of this history is not that combatants should switch to guns. The much-enhanced technology of gunfire today, including the easy firing of multiple rounds, makes that option ever less socially palatable, if it ever was.

Instead, the clear requirement is to separate combatants and to ritualise the expression of social and personal aggression. Achieving such policies must rely considerably upon systems of law and policing. Yet socio-cultural attitudes among the wider public are highly relevant too. As the history of duelling indicates, even august Prime Ministers allowed themselves upon occasion to be provoked into behaving in ways that put them at risk of criminal charges. But changing social mores eventually removed that option, even for the most combative and headstrong of politicians today. Community attitudes at first ritualised the personal resolution of conflicts and eventually withdrew support for such behaviour entirely.

So today multiple approaches are required. Police actions to discourage young men from carrying knives constitute an obvious and important step. Ditto effective policies to curb the drug culture. Equally crucial are strong and repeated expressions of community disapproval of violence and knife-carrying. Yet policing and public attitudes can’t work without complementary interventions to combat youth alienation and, especially, to provide popular non-violent outlets for energy and aggression. Leaving bored young people feeling fearful and at risk in public places is no recipe for social order.

How can energies and aggression be either ritualised and/or channelled into other outlets? It’s for young people and community activists to specify. But many potential options spring to mind: youth clubs; youth theatre; participatory sports of all kinds; martial arts; adventure programmes; community and ecological projects; music-making festivals; dance; creative arts; church groups; … let alone continuing educational access via further education study grants. It’s true that all such plans involve constructive imagination, organisation, and expenditure. But their benefits are immense. Violence happens within societies; and so, very emphatically, does conflict resolution and, better still, the redirection of energies and aggression into constructive pathways.

1 See variously S. Banks, Duels and Duelling (Oxford, 2014); U. Frevert, Men of Honour: A Social and Cultural History of the Duel (Cambridge, 1995); V.G. Kiernan, The Duel in European History: Honour and the Reign of the Aristocracy (Oxford, 1988; 2016); M. Peltonen, The Duel in Early Modern England: Civility, Politeness and Honour (Cambridge, 2003); P. Spierenburg (ed.), Men and Violence: Gender, Honour and Rituals in Modern Europe and America (Columbus, Ohio, 1998).

2 S. Banks, ‘Dangerous Friends: The Second and the Later English Duel’, Journal of Eighteenth-Century Studies, 32 (2009), pp. 87-106.

3 R.G. Shoemaker, ‘The Taming of the Duel: Masculinity, Honour and Ritual Violence in London, 1660-1800’, Historical Journal, 45 (2002), pp. 525-45.

4 S. Richardson, The History of Sir Charles Grandison (1753; in Oxford 1986 edn), Bk.1, pp. 207-8.

5 B. Franklin, ‘On Duelling’ (1784), in R.L. Ketcham (ed.), The Political Thought of Benjamin Franklin (Indianapolis, Ind., 1965; 2003), p. 362. For context, see also W.O. Stevens, Pistols at Ten Paces: The Story of the Code of Honour in America (Boston, 1940); D. Steward, Duels and the Roots of Violence in Missouri (2000); and C. Burchfield, Choose Your Weapon: The Duel in California, 1847-61 (Fresno, CA., 2016).

6 R.B. Sheridan, The Rivals (1775), ed E. Duthie (1979), Act V, sc. 2 + 3, pp. 105, 112. For the Irish context, see J. Kelly, ‘That Damn’d Thing Called Honour’: Duelling in Ireland, 1570-1860 (Cork, 1995).

For further discussion, see

To read other discussion-points, please click here

To download Monthly Blog 100 please click here

MONTHLY BLOG 97, WHY IS THE REMARKABLE CHARLOTTE DESPARD NOT BETTER KNOWN?

If citing, please kindly acknowledge copyright © Penelope J. Corfield (2019)

Fig.1 Charlotte Despard speaking at an anti-fascist rally, Trafalgar Square, 12 June 1933:
photograph by James Jarché, Daily Herald Archive.

Charlotte Despard (1844-1939) was a remarkable – even amazing – woman. Don’t just take my word for it. Listen to Mahatma Gandhi (1869-1948). Visiting London in 1909, he met all the leading suffragettes. The one who impressed him most was Charlotte Despard. She is ‘a wonderful person’, he recorded. ‘I had long talks with her and admire her greatly’.1 They both affirmed their faith in the non-violent strategy of political protest by civil disobedience. Despard called it ‘spiritual resistance’.

What’s more, non-violent protest has become one of the twentieth-century’s greatest contributions to potent mass campaigning – without resorting to counter-productive violence. Associated with this strategy, the names of Henry Thoreau, Mahatma Gandhi and Martin Luther King, all controversial in their day, have become canonised.2 Yet Charlotte Despard, who was also controversial in her day, has been substantially dropped from the historical record.

Not entirely so. On 14 December 2018 Battersea Labour unveiled a blue plaque in her honour, exactly one hundred years after the date when she stood as the Labour Party candidate in North Battersea in the 1918 general election. She was one of the feminist pioneers, when no more than sixteen women stood. But Despard lost heavily to the Liberal candidate, even though industrial North Battersea was then emerging as a Labour stronghold.3

And one major reason for her loss helps to explain her disappearance from mainstream historical memory. Despard was a pacifist, who opposed the First World War and campaigned against conscription. Many patriotic voters in Battersea disagreed with this stance. In the immediate aftermath of war, emotions of relief and pride triumphed. Some months later, Labour swept the board in the 1919 Battersea municipal elections; but without Charlotte Despard on the slate.

Leading pacifists are not necessarily all neglected by history.4 But the really key point was that Charlotte Despard campaigned for many varied causes during her long life and, at every stage, weakened her links with previous supporters. Her radical trajectory made complete sense to her. She sought to befriend lame dogs and to champion outsiders. Yet as an independent spirit – and seemingly a psychological loner – she walked her own pathway.

Despard was by birth an upper crust lady of impeccable Anglo-Irish ancestry, with high-ranking military connections. For 40 years, she lived quietly, achieving a happy marriage and a career as a minor novelist. Yet, after being widowed at the age of 40, she had an extraordinary mid- and late-life flowering. She moved to Battersea’s Nine Elms, living among the poorest of the poor. And she then became a life-long radical campaigner. By the end of her career, she was penniless, having given all her funds to her chosen causes.

A convinced suffragette, Despard joined the Women’s Social and Political Union and was twice imprisoned for her public protests. In 1907, however, she was one of the leading figures to challenge the authoritarian leadership style of Christabel Pankhurst. Despard resigned and founded the rival Women’s Freedom League. This smaller group opposed the use of violence. Instead, its members took symbolic action, like unfurling banners in parliament. They also advocated passive resistance, like non-payment of taxation and non-cooperation with the census. (I recently discovered, thanks to the research of a family member, that my great-grandmother was a would-be WFL supporter. So the 1911 census enumerator duly noted that Mrs Matilda Corfield, living in Sheffield, had given information only ‘under Protest (she wants the vote)’.5 This particular example of resistance was very muffled and inconsequential. Nevertheless, it indicated how unknown women across the country tried to respond to WFL advice. It was one way of slowly changing the climate of public opinion.)

However, the energetic Charlotte Despard did not confine her efforts solely to the cause of the female suffrage. Her life in Battersea radicalised her politically and she became a socialist. She was not good at detailed committee work. Her forte was activism. Indefatigably, she organised a local welfare system. She funded health centres for mothers and babies, exchange points for cots and equipment, youth clubs, and halls for local meetings. And the front room of her small premises in Nine Elms was made available to the public as a free reading room, stocked with books and newspapers. It was a one-woman exercise in practical philanthropy. What’s more, her 1918 election manifesto called for a minimum wage – something not achieved until 1998.

Among the Battersea workers, the tall, wiry, and invariably dignified Charlotte Despard cut an impressive figure. A lifelong vegetarian, she was always active and energetic. And she believed in the symbolic importance of dress. Thus she habitually wore sandals (or boots in winter) under long, flowing robes, a lace shawl, and a mantilla-like head-dress. The result was a timeless style, unconcerned with passing fashions. She looked like a secular sister of mercy.
2019-01-No2-Charlotte-Despard-in-slumland

Fig.2 Charlotte Despard in the poor tenements of Battersea’s Nine Elms, where she lived from 1890 to the early 1920s, instituting and funding local welfare services. Her visitors commented adversely on the notorious ‘Battersea smell’ of combined industrial effluent and smoke from innumerable coalfires; but Despard reportedly took no notice.

For a number of years, Despard worked closely with the newly founded Battersea Labour Party (1908- ), strengthening its global connections. She attended various international congresses; and she backed the Indian communist Shapurji Saklatvala as the Labour-endorsed candidate in Battersea North at the general election in 1922. (He won, receiving over 11,000 votes). Yet, as already noted, the Battersea electorate in 1918 had rebuffed her own campaign.

Then at a relatively loose end, Despard moved to Dublin in the early 1920s. She had already rejected her Irish Ascendancy background by converting to Catholicism. There she actively embraced the cause of Irish nationalism and republicanism. She became a close supporter of Maud Gonne, the charismatic exponent of Irish cultural and political independence. By the later 1920s, however, Despard was unhappy with the conservatism of Irish politics. In 1927 she was classed as a dangerous subversive by the Free State, for opposing the Anglo-Irish Treaty settlement. She eventually moved to Belfast and changed tack politically to endorse Soviet communism. She toured Russia and became secretary of the British Friends of the Soviet Union (FSU), which was affiliated to the International Organisation of the same name.

During this variegated trajectory, Despard in turn shocked middle-class suffragettes who disliked her socialism. She then offended Battersea workers who rejected her pacifism. She next infuriated English Protestants who hated her Irish nationalism. And she finally outraged Irish Catholics (and many Protestants as well) who opposed her support for Russian communism. In 1933, indeed, her Dublin house was torched and looted by an angry crowd of Irish anti-communists.6

In fact, Despard always had her personal supporters, as well as plentiful opponents. But she did not have one consistent following. She wrote no autobiography; no memorable tract of political theory. And she had no close family supporters to tend her memory. She remained on good terms with her younger brother throughout her life. But he was Sir John French, a leading military commander in the British Army and from 1918 onwards Lord Lieutenant of Ireland. The siblings disagreed politically on everything – although both shared the capacity to communicate on easy terms with people from many different backgrounds. To the Despards, ‘Aunt Lottie’ was thus an eccentric oddity. To other respectable family friends, she was ‘a witch’, and a dangerous one at that.7

These factors combined together to isolate Despard and to push her, after her death, into historical limbo. There are very few public monuments or memorials to her indomitable career. In north London, a pleasant pub on the Archway Road is named after her, on land which was owned by her husband Colonel Despard. On Battersea’s Doddington Estate, there is an avenue named after her, commemorating her welfare work in the area. And now there is the blue plaque outside the headquarters of Battersea Labour at 177 Lavender Hill, SW11. These memorials are fine but hardly enough.

Fig.3 Blue plaque to Charlotte Despard, outside 177 Lavender Hill, London SW11 5TE: installed 14 December 2018, on the precise centenary of her standing for parliament in 1918, as one of only 16 women pioneers to do so.

Why should she be remembered? The answer is not that everyone would have agreed (then or later) with all of Charlotte Despard’s political calls. As this account has shown, she was always controversial and, on Russia, self-deceived into thinking it much more of a workers’ paradise than it was (as were many though not all left-leaning intellectuals in the West). Nonetheless, she is a remarkable figure in the history of public feminism. She not only had views but she campaigned for them, using her combination of practical on-the-ground organisation, her call for symbolic non-violent protest and ‘spiritual resistance’, and her public oratory. And she did so for nigh on 50 years into her very advanced years.

Indomitability, peaceful but forceful, was her signature style. She quoted Shelley on the need for Love, Hope, and Endurance. When she was in her mid-sixties, she addressed a mass rally in Trafalgar Square (of course, then without a microphone). Her speeches were reportedly allusive and wide-ranging, seeking to convey inspiration and urgency. One onlooker remembered that her ‘thin, fragile body seemed to vibrate with a prophecy’.8

Appropriately for a radical campaigner, Charlotte Despard’s last major public appearance was on 12 June 1933, when she spoke passionately at a mass anti-fascist rally in Trafalgar Square. At that time, she was aged 89. It was still unusual then for women to speak out boldly in public. They often faced jeers and taunts for doing so. But the photographs of her public appearances show her as unflinching, even when she was the only woman amidst crowds of men. Above all, for the feminist feat of speaking at the mass anti-fascist rally at the age of 89, there is a good case for placing a statue on Trafalgar Square’s vacant fourth plinth, showing Despard in full oratorical flow. After all, she really was there. And, if not on that particular spot, then somewhere relevant in Battersea. Charlotte Despard, born 175 years ago and campaigning up until the start of the Second World War, was a remarkable phenomenon. Her civic and feminist commitment deserves public commemoration – and in a symbolic style worthy of the woman.

Figs 4 + 5: Photos showing Despard, speaking in Trafalgar Square, without a microphone:
(L) dated 1910 when she was 66, and (R) dated 1933 when she was aged 89.
Her stance and demeanour are identically rapt, justifying one listener’s appreciative remark:
Mrs Despard – she always gets a crowd’.

1 Quoted in M. Mulvihill, Charlotte Despard: A Biography (1989), p. 86. See also A. Linklater, An Unhusbanded Life: Charlotte Despard, Suffragette, Socialist and Sinn Feiner (1980); and, for Battersea context, P.J. Corfield in Battersea Matters (Autumn 2016), p. 11; and PJC with Mike Marchant, DVD: Red Battersea: One Hundred Years of Labour, 1908-2008 (2008).

2 A. Roberts and T. Garton Ash (eds), Civil Resistance and Power Politics: The Experience of Non-Violent Action from Gandhi to the Present (Oxford, 2009); R.L. Holmes and B.L. Gan (eds), Nonviolence in Theory and Practice (Long Grove, Illinois, 2012).

3 1918 general election result for North Battersea: Richard Morris, Liberal (11,231 = 66.6% of all voting); Charlotte Despard, Labour (5,634 = 33.4%). Turnout =  43.7%.

4 P. Brock and N. Young, Pacifism in the Twentieth Century (New York, 1999).

5 With thanks to research undertaken by Annette Aseriau.

6 Mulvihill, Charlotte Despard, p. 180.

7 Ibid., pp. 46-7, 78-9.

8 Account by Christopher St John, in Mulvihill, Charlotte Despard, p. 77.

For further discussion, see

To read other discussion-points, please click here

To download Monthly Blog 97 please click here

MONTHLY BLOG 96, WHAT’S WRONG WITH PREHISTORY?

If citing, please kindly acknowledge copyright © Penelope J. Corfield (2018)

Arthur’s Stone, Herefordshire, dating from c.3000 BCE: photo © Tony Belton, 2016

Arthur’s Stone, Herefordshire, dating from c.3000 BCE:
photo © Tony Belton, 2016

What’s wrong with ‘prehistory’? Absolutely nothing but the name. People refer to ancient monuments as ‘prehistoric’ and everyone knows roughly what is meant. The illustration (above) shows an ancient burial tomb, known as Arthur’s Stone, dating from 3000 BCE, which I visited in Herefordshire on a summer day in 2016. It did and does indeed look truly venerable. So loose terms such as ‘prehistoric’ are passable enough if used casually.

But ‘prehistory’ as a scholarly term in application to a prolonged period of human history? Seriously misleading. It implies that the long aeons of foundational human history, before the advent of literacy, somehow occurred in a separate ante-chamber to the ‘real’ deal.

The acquiring of skills in reading and writing (which occurred in different parts of the world at different times) was in fact part of a lengthy process of human adaptation and invention. Before literacy, key developments included: the adoption of clothing; the taming of fire; the invention of tools; the refinement of tools and weapons with handles; the invention of the wheel; the arrival of speech; the advent of decorative arts; the formulation of burial rituals; the domestication of animals; the development of a calendrical consciousness; the capacity to cope with population fluctuations including survival during the Ice Age; the start of permanent settlements and farming; and the cumulative mental and cultural preparation for the invention of reading and writing. Some list! The pace of change was often slow; but the changes were absolutely foundational to human history.1

In practice, of course, the skilled and ingenious experts, who study pre-literate societies, do not consider their subject to be anything other than fully and deeply historical. They use ‘prehistory’ because it is a known term of art. (Often, indeed, they may start their lectures and books with a jovial disclaimer that such terminology should not be taken literally). The idea of ‘prehistory’ was crystallised by Victorian historians, who were developing a deep reverence for the importance of written sources for writing ‘real’ history. But the differences in prime source material, although methodologically significant, are not fundamental enough to deprive the foundational early years of the full status of history. And, in fact, these days historians of all periods study a range of sources. They are not just stuck in archives, reading documents – important as those are. If relevant to their theme, historians may examine buildings, art, artefacts, materials, bones, refuse, carbon datings, statistical extrapolations, and/or genetic evidence (etc etc), just as do archaeologists and ‘prehistorians’.

Moreover, conventional references to ‘prehistory’ have now been blind-sided by the recent return to diachronic (through-time) studies of what is known as Big History. This approach to the past takes as its remit either the whole of the cosmos or at least the whole lifespan of Planet Earth.2 It draws upon insights from cosmologists and astro-physicists, as well as from geologists and biologists. After all, a lot of history had indeed happened before the first humans began to walk. So what are the millennia before the advent of homo sapiens to be entitled? Pre-prehistory? Surely not. All these eras form part of what is sometimes known as ‘deep history’: a long time ago but still historical.

So why has the misleading term ‘prehistory’ survived for so long? One major reason lies in the force of inertia – or institutional continuity, to give it a kinder name. ‘Prehistory’ has prevailed as an academic terminology for over a century. It appears in the names of academic departments, research institutions, learned societies, job descriptions, teaching courses, examination papers, academic journals, books, blogs, conferences, publishers’ preferences for book titles, and popular usages – let alone in scholars’ self-definitions. Little wonder that renaming is not a simple matter. Nonetheless, subjects are continuously being updated – so why not a further step now?

I was prompted to write on this question when three congenial colleagues asked me, a couple of years ago, to contribute to a volume on Time & History in Prehistory (now available, with publication date 2019).3 I was keen to respond but hostile to the last word in their book title. My answer took the form of arguing that this specialist section of historical studies needs a new and better name. I am grateful to the editors’ forbearance in accepting my contribution. It contributes to debates elsewhere within the volume, since criticising the terminology of ‘prehistory’ is not new.

Apart from the lack of logic in apparently excluding the foundational experiences of the human species from ‘real’ history, my own further objection is that the division inhibits diachronic analysis of the long term. A surviving relic from ‘prehistoric’ times, like Arthur’s Stone, has a long and intriguing history which still continues. At some stage long before the thirteenth century CE, the modest monument, high on a ridge between the Wye and Golden Valleys, became associated in popular legend with the feats of King Arthur. (Did he win a battle there, rumour speculated, or slay a giant?) That invented linkage is in itself a fascinating example of the spread of the Arthurian legend.4

The site later witnessed some real-life dramas. In the fifteenth century, a knight was killed there in a fatal duel. And in September 1645 the embattled Charles I dined at the Stone with his royalist troops. Perhaps he intended the occasion as a symbolic gesture, although it did not confer upon him sufficient pseudo-Arthurian lustre to defeat Cromwell and the Roundheads.

For the villagers in nearby Dorstone and Bredwardine, Arthur’s Stone at some stage (chronology uncertain) became a venue for popular festivities, with dancing and ‘high jinks’ every midsummer. This long-standing tradition continued until well into Victorian times. As a sober counter-balance, too, the local Baptists in the nineteenth and twentieth centuries organised an ecumenical religious service there each June/July. Living witnesses remember these as occasions of fervent al fresco hymn-singing. Implicitly, they were acknowledging the Stone’s sacral nature, whilst simultaneously purging its pagan associations.

When visiting the Stone myself in 2016, I met by chance a local resident, named Ionwen Williams. In a stroke of research serendipity, we got chatting and she zestfully recounted her memories, as a child before World War II, of joining her schoolfellows to sing hymns at the site each midsummer. This experience and many later visits confirmed for her the special nature of the place. I did not for a moment doubt her memories; but, as a prudent historian, thought it helpful to cross-check – and found them corroborated.

It is abundantly clear that, throughout its five thousand years of existence, Arthur’s Stone has had multiple meanings for the witnessing generations. At one sad stage in the late nineteenth century, it was pillaged by builders taking stones for new constructions. But local objections put a stop to that; and it is now guarded by English Heritage. It is utterly historic, not separately ‘prehistoric’: and the same point applies to all long-surviving monuments, many of which are much bigger and more famous than Arthur’s Stone. Furthermore, deep continuities apply to many other aspects of human history – and not just to physical monuments. For example, there are many claims and counter-claims about the foundations of human behaviour which merit debate, without compartmentalising the eras of pre-literacy from those of post-literacy.

Lastly, what alternative nomenclature might apply? Having in the first draft of my essay rebuked the specialists known as ‘prehistorians’ for not changing their name, I was challenged by the editors to review other options. Obviously it’s not for one individual to decide. It was, however, a good challenge. In many ways, these early millennia might be termed ‘foundational’ in human history. That, after all, is what they were. On the other hand, ‘foundational history’ sounds like a first-year introduction course. Worthy but not very evocative. My essay reviews various options and plumps for ‘primeval’ history. That term not only sounds ancient but signals primacy: in human history, these years came first.5 The contributions within the volume as a whole are questioning and challenging throughout, as they analyse different aspects of Time and, yes, ‘History’. It is a pleasure to join these essays in thinking long.6

1 For an enticing introduction (apart from one word in its subtitle), see C. Gamble, Timewalkers: The Prehistory of Global Colonisation (Sutton: Stroud 1993).

2 For an introduction, see D.G. Christian, Maps of Time: An Introduction to Big History (U. of California Press: Berkeley, 2004).

3 S. Souvatzi, A. Baysal and E.L. Baysal (eds), Time and History in Prehistory (Routledge: Abingdon, 2019).

4 N.J. Lacy (ed.), The New Arthurian Encyclopaedia (Garland: New York, 1991).

5 P.J. Corfield, ‘Primevalism: Saluting a Renamed Prehistory’, in Soutvatzi, Baysal and Baysal (eds), Time and History, pp. 265-82. My own interest in ‘long ago’ was sparked when, as a teenager, I read a study by Ivar Lissner, entitled The Living Past (Cape: London, 1957): for which see P.J. Corfield, ‘An Unknown Book Which Influenced Me’ BLOG no.14 (Nov. 2011).

6 On this theme, see J. Guldi and D. Armstrong, The History Manifesto (Cambridge University Press: Cambridge, 2014); P.J. Corfield, ‘What on Earth is the “Temporal Turn” and Why is it Happening Now?’ BLOG no.49 (Jan. 2015); and idem, ‘Thinking Long: Studying History’, BLOG no.94 (Oct. 2018), all BLOGs available on www.penelopejcorfield.com/monthly-blogs.

For further discussion, see

To read other discussion-points, please click here

To download Monthly Blog 96 please click here

MONTHLY BLOG 95, ‘WHAT IS THE GREATEST SIN IN THE WORLD?’ CHRISTOPHER HILL AND THE SPIRIT OF EQUALITY

If citing, please kindly acknowledge copyright © Penelope J. Corfield (2018)

Text of short talk given by PJC to introduce the First Christopher Hill Memorial Lecture, (given by Prof. Justin Champion) at Newark National Civil War Centre, on Saturday 3 November 2018.

Christopher Hill was not only a remarkable historian – he was also a remarkable person.1 All his life, he believed, simply and staunchly, in human equality. But he didn’t parade his beliefs on his sleeve. At first meeting, you would have found him a very reserved, very solid citizen. And that’s because he was very reserved – and he was solid in the best sense of that term. He was of medium height, so did not tower over the crowd. But he held himself very erect; had a notably sturdy, broad-shouldered Yorkshire frame; and was very fit, cycling and walking everywhere. And in particular, Christopher Hill had a noble head, with a high forehead, quizzical eyebrows, and dark hair which rose almost vertically – giving him, especially in his later years, the look of a wise owl.
Christopher-Hill-1-&-2

Christopher Hill (L) in his thirties and (R) in his seventies

By the way, he was not a flashy dresser. The Hill family motto was ‘No fuss’. And, if you compare the two portraits of him in his 30s and his 70s, you could be forgiven for thinking that he was wearing the same grey twill jacket in both. (He wasn’t; but he certainly stuck to the same style all his life).

Yet even while Christopher Hill was reserved and dignified, he was also a benign figure. He had no side. He did not pull rank. He did not demand star treatment. He was courteous to all – and always interested in what others had to say. That was a key point. As Master of Balliol, Hill gave famous parties, at which dons and students mingled; and he was often at the centre of a witty crowd. But just as much, he might be found in a corner of the room discussing the problems of the world with a shy unknown.

As I’ve already said. Christopher Hill believed absolutely in the spirit of equality. But he did know that it was a hard thing to achieve – and that was why he loved the radicals in the English civil wars of the mid-seventeenth century. They were outsiders who sought new ways of organising politics and religion. Indeed, they struggled not only to define equality – but to live it. And, although there was sometimes a comic side to their actions, he admired their efforts.

When I refer to unintentionally comic aspects, I am thinking of those Ranters, from the radical and distinctly inchoate religious group, who jumped up in church and threw off their clothes as a sign. The sign was that they were all God’s children, equal in a state of nature. Not surprisingly, such behaviour attracted a lot of criticism – and satirists had good fun at their expense.

Well, Christopher Hill was far too dignified to go around throwing off his clothes. But he grew up believing a radical form of Methodism, which stressed that ‘we are all one in the eyes of the Lord’. As I’ve said, his egalitarianism came from within. But he was clearly influenced by his Methodist upbringing. His parents were kindly people, who lived simply and modestly (neither too richly nor too poorly). They didn’t drink, didn’t smoke, didn’t swear and didn’t make whoopee. Twice and sometimes even three times on Sundays, they rode their bikes for several miles to and from York’s Central Methodist Chapel; and then discussed the sermon over lunch.

In his mid-teens, Hill was particularly inspired by a radical Methodist preacher. He was named T.S. Gregory and he urged a passionate spiritual egalitarianism. Years later, Hill reproduced for me Gregory’s dramatic pulpit style. He almost threw himself across the lectern and spoke with great emphasis: ‘Go out into the streets – and look into the eyes of every fellow sinner, even the poorest beggar or the most abandoned prostitute; [today he would add look under the hoods of the druggies and youth gangs]; look into these outcast faces and in every individual you will see elements of the divine. The York Methodists, from respectable middle class backgrounds, were nonplussed. But Hill was deeply stirred. For him, Gregory voiced a true Protestantism – which Hill defined as wine in contrast with what he saw as the vinegar and negativism of later Puritanism.

The influence of Gregory was, however, not enough to prevent Hill in his late teens from losing his religious faith. My mother, Christopher’s younger sister, was very pleased at this news as she welcomed his reinforcement. She herself had never believed in God, even though she too went regularly to chapel. But their parents were sincerely grieved. On one occasion, there was a dreadful family scene, when Christopher, on vacation from Oxford University, took his younger sister to the York theatre. Neither he nor my mother could later remember the show. But they both vividly recalled their parent’s horror: going to the theatre – abode of the devil! Not that the senior Hills shouted or rowed. That was not their way. But they conveyed their consternation in total silence … which was difficult for them all to overcome.

As he lost his faith, Hill converted to a secular philosophy, which had some elements of a religion to it. That was Marxism. Accordingly, he joined the British Communist Party. And he never wavered in his commitment to a broad-based humanist Marxism, even when he resigned from the CP in 1956. Hill was not at all interested in the ceremonies and ritual of religion. The attraction of Marxism for him was its overall philosophy. He was convinced that the revolutionary unfolding of history would eventually remove injustices in this world and usher in true equality. Hill sought what we would call a ‘holistic vision’. But the mover of change was now History rather than God.

On those grounds, Hill for many years supported Russian communism as the lead force in the unfolding of History. In 1956, however, the Soviet invasion of Hungary heightened a fierce internal debate within the British Communist Party. Hill and a number of his fellow Marxist historians, struggled to democratise the CP. But they lost and most of them thereupon resigned.

This outcome was a major blow to Hill. Twice he had committed to a unifying faith and twice he found its worldly embodiment unworthy. Soviet Communism had turned from intellectual inspiration into a system based upon gulags, torture and terror. Hill never regretted his support for Soviet Russia during the Second World War; but he did later admit that, afterwards, he had supported Stalinism for too long. The mid-1950s was an unhappy time for him both politically and personally. But, publicly, he did not wail or beat his breast. Again, that was not the Hill way.

He did not move across the political spectrum, as some former communists did, to espouse right-wing causes. Nor did he become disillusioned or bitter. Nor indeed, did he drop everything to go and join a commune. Instead, Hill concentrated even more upon his teaching and writing. He did actually join the Labour Party. Yet, as you can imagine, his heart was not really in it.

It was through his historical writings, therefore, that Hill ultimately explored the dilemmas of how humans could live together in a spirit of equality. The seventeenth-century conflicts were for him seminal. Hill did not seek to warp history to fit his views. He could not make the radicals win, when they didn’t. But he celebrated their struggles. For Hill, the seventeenth-century religious arguments were not arid but were evidence of the sincere quest to read God’s message. He had once tried to do that himself. And the seventeenth-century political contests were equally vivid for him, as he too had been part of an organised movement which had struggled to embody the momentum of history.

As I say, twice his confidence in the worldly formulations of his cause failed. Yet his belief in egalitarianism did not. Personally, he became happy in his second marriage; and he immersed himself in his work as a historian. From being a scholar who wrote little, he became super-productive. Books and essays poured from his pen. Among those he studied was the one seventeenth-century radical who appealed to him above all others: Gerrard Winstanley, the Digger, who founded an agrarian commune in the Surrey hills. And the passage in Winstanley’s Law of Freedom (1652) that Hill loved best was dramatic in the best T.S. Gregory style. What is the greatest sin in the world? demanded Winstanley. And he answered emphatically that it is for rich people to hoard gold and silver, while poor people suffer from hunger and want.          

          What Hill would say today, at the ever widening inequalities across the world, is not hard to guess. But he would also say: don’t lose faith in the spirit of equality. It is a basic tenet of human life. And all who believe in fair does for all, as part of true freedom, should strive to find our own best way, individually and/or collectively, to do our best for our fellow humans and to advance Hill’s Good Old Cause.

1 For documentation, see P.J. Corfield, ‘“We are all One in the Eyes of the Lord”: Christopher Hill and the Historical Meanings of Radical Religion’, History Workshop Journal, 58 (2004), pp. 110-27. Now posted on PJC personal website as Pdf5; and further web-posted essays PJC Pdf47-50, all on www.penelopejcorfield.co.uk

For further discussion, see

To read other discussion-points, please click here

To download Monthly Blog 95 please click here

MONTHLY BLOG 94, THINKING LONG – STUDYING HISTORY

If citing, please kindly acknowledge copyright © Penelope J. Corfield (2018)

History is a subject that deals in ‘thinking long’. The human capacity to think beyond the immediate instant is one of our species’ most defining characteristics. Of course, we live in every passing moment. But we also cast our minds, retrospectively and prospectively, along the thought-lines of Time, as we mull over the past and try to anticipate the future. It’s called ‘thinking long’.

Studying History (indicating the field of study with a capital H) is one key way to cultivate this capacity. Broadly speaking, historians focus upon the effects of unfolding Time. In detail, they usually specialise upon some special historical period or theme. Yet everything is potentially open to their investigations.

Sometimes indeed the name of ‘History’ is invoked as if it constitutes an all-seeing recording angel. So a controversial individual in the public eye, fearing that his or her reputation is under a cloud, may proudly assert that ‘History will be my judge’. Quite a few have made such claims. They express a blend of defiance and  optimism. Google: ‘History will justify me’ and a range of politicians, starting with Fidel Castro in 1963, come into view. However, there’s no guarantee that the long-term verdicts will be kinder than any short-term criticisms.

True, there are individuals whose reputations have risen dramatically over the centuries. The poet, painter and engraver William Blake (1757-1827), virtually unknown in his own lifetime, is a pre-eminent example. Yet the process can happen in reverse. So there are plenty of people, much praised at the start of their careers, whose reputations have subsequently nose-dived and continue that way. For example, some recent British Prime Ministers may fall into that category. Only Time (and the disputatious historians) will tell.

Fig. 1 William Blake’s Recording Angel has about him a faint air of an impish magician as he points to the last judgment. If this task were given to historians, there would be a panel of them, arguing amongst themselves.

In general, needless to say, those studying the subject of History do not define their tasks in such lofty or angelic terms. Their discipline is distinctly terrestrial and Time-bound. It is prone to continual revision and also to protracted debates, which may be renewed across generations. There’s no guarantee of unanimity. One old academic anecdote imagines the departmental head answering the phone with the majestic words: ‘History speaking’.1 These days, however, callers are likely to get no more than a tinny recorded message from a harassed administrator. And academic historians in the UK today are themselves being harried not to announce god-like verdicts but to publish quickly, in order to produce the required number of ‘units of output’ (in the assessors’ unlovely jargon) in a required span of time.

Nonetheless, because the remit of History is potentially so vast, practitioners and students have unlimited choices. As already noted, anything that has happened within unfolding Time is potentially grist to the mill. The subject resembles an exploding galaxy – or, rather, like the cosmos, the sum of many exploding galaxies.

Tempted by that analogy, some practitioners of Big History (a long-span approach to History which means what it says) do take the entire universe as their remit, while others stick merely to the history of Planet Earth.2 Either way, such grand approaches are undeniably exciting. They require historians to incorporate perspectives from a dazzling range of other disciplines (like astro-physics) which also study the fate of the cosmos. Thus Big History is one approach to the subject which very consciously encourages people to ‘think long’. Its analysis needs careful treatment to avoid being too sweeping and too schematic chronologically, as the millennia rush past. But, in conjunction with shorter in-depth studies, Big History gives advanced students a definite sense of temporal sweep.

Meanwhile, it’s also possible to produce longitudinal studies that cover one impersonal theme, without having to embrace everything. Thus there are stimulating general histories of the weather,3 as well as more detailed histories of weather forecasting, and/or of changing human attitudes to weather. Another overarching strand studies the history of all the different branches of knowledge that have been devised by humans. One of my favourites in this genre is entitled: From Five Fingers to Infinity.4 It’s a probing history of mathematics. Expert practitioners in this field usually stress that their subject is entirely ahistorical. Nonetheless, the fascinating evolution of mathematics throughout the human past to become one globally-adopted (non-verbal) language of communication should, in my view, be a theme to be incorporated into all advanced courses. Such a move would encourage debates over past changes and potential future developments too.

Overall, however, the great majority of historians and their courses in History take a closer focus than the entire span of unfolding Time. And it’s right that the subject should combine in-depth studies alongside longitudinal surveys. The conjunction of the two provides a mixture of perspectives that help to render intelligible the human past. Does that latter phrase suffice as a summary definition?5 Most historians would claim to study the human past rather than the entire cosmos.

Yet actually that common phrase does need further refinement. Some aspects of the human past – the evolving human body, for example, or human genetics – are delegated for study by specialist biologists, anatomists, geneticists, and so forth. So it’s clearer to say that most historians focus primarily upon the past of human societies in the round (ie. including everything from politics to religion, from war to economics, from illness to health, etc etc). And that suffices as a definition, provided that insights from adjacent disciplines are freely incorporated into their accounts, wherever relevant. For example, big cross-generational studies by geneticists are throwing dramatic new light upon the history of human migration around the globe and also of intermarriage within the complex range of human species and the so-called separate ‘races’ within them.6 Their evidence amply demonstrates the power of longitudinal studies for unlocking both historical and current trends.

The upshot is that the subject of History can cover everything within the cosmos; that it usually concentrates upon the past of human societies, viewed in the round; and that it encourages the essential human capacity for thinking long. For that reason, it’s a study for everyone. And since all people themselves constitute living histories, they all have a head-start in thinking through Time.7

1 I’ve heard this story recounted of a formidable female Head of History at the former Bedford College, London University; and the joke is also associated with Professor Welch, the unimpressive senior historian in Kingsley Amis’s Lucky Jim: A Novel (1953), although upon a quick rereading today I can’t find the exact reference.

2 For details, see the website of the Big History’s international learned society (founded 2010): www.ibhanet.org. My own study of Time and the Shape of History (2007) is another example of Big History, which, however, proceeds not chronologically but thematically.

3 E.g. E. Durschmied, The Weather Factor: How Nature has Changed History (2000); L. Lee, Blame It on the Rain: How the Weather has Changed History (New York, 2009).

4 F.J. Swetz (ed.), From Five Fingers to Infinity: A Journey through the History of Mathematics (Chicago, 1994).

5 For meditations on this theme, see variously E.H. Carr, What is History? (Cambridge 1961; and many later edns); M. Bloch, The Historian’s Craft (in French, 1949; in English transl. 1953); B. Southgate, Why Bother with History? Ancient, Modern and Postmodern Motivations (Harlow, 2000); J. Tosh (ed.), Historians on History: An Anthology (2000; 2017); J. Black and D.M. MacRaild, Studying History (Basingstoke, 2007); H.P.R. Finberg (ed.), Approaches to History: A Symposium (2016).

6 See esp. L.L. Cavalli-Sforza and F. Cavalli-Sforza, The Great Human Diasporas: The History of Diversity and Evolution, transl. by S. Thomas (Reading, Mass., 1995); D. Reich, Who We Are and Where We Got Here: Ancient DNA and the New Science of the Human Past (Oxford, 2018).

7 P.J. Corfield, ‘All People are Living Histories: Which is why History Matters’. A conversation-piece for those who ask: Why Study History? (2008) in London University’s Institute of Historical Research Project, Making History: The Discipline in Perspective www.history.ac.uk/makinghistory/resources/articles/why_history_matters.html; and also available on www.penelopejcorfield.co.uk/ Pdf1.

For further discussion, see

To read other discussion-points, please click here

To download Monthly Blog 94 please click here

MONTHLY BLOG 92, HISTORIANS AT WORK THROUGH TIME

If citing, please kindly acknowledge copyright © Penelope J. Corfield (2018)
Historians, who study the past, don’t undertake this exercise from some vantage point outside Time. They, like everyone else, live within an unfolding temporality. That’s very fundamental. Thus it’s axiomatic that historians, like their subjects of study, are all equally Time-bound.1

Nor do historians undertake the study of the past in one single moment in time. Postmodernist critics of historical studies sometimes write as though historical sources are culled once only from an archive and then adopted uncritically. The implied research process is one of plucking choice flowers and then pressing them into a scrap-book to some pre-set design.

On such grounds, critics of the discipline highlight the potential flaws in all historical studies. Sources from the past are biased, fallible and scrappy. Historians in their retrospective analysis are also biased, fallible and sometimes scrappy. And historical writings are literary creations only just short of pure fiction.2

Historians should welcome scepticism this dose of scepticism – always a useful corrective. Yet they entirely reject the proposition that trying to understand bygone eras is either impossible or worthless. Rebuttals to postmodernist scepticism have been expressed theoretically;3 and also directly, via pertinent case studies which cut through the myths and ‘fake news’ which often surround controversial events in history.4

When at work, historians should never take their myriad of source materials literally and uncritically. Evidence is constantly sought, interrogated, checked, cross-checked, compared and contrasted, as required for each particular research theme. The net is thrown widely or narrowly, again depending upon the subject. Everything is a potential source, from archival documents to art, architecture, artefacts and though the gamut to witness statements and zoological exhibits. Visual materials can be incorporated either as primary sources in their own right, or as supporting documentation. Information may be mapped and/or tabulated and/or statistically interrogated. Digitised records allow the easy selection of specific cases and/or the not-so-easy processing of mass data.

As a result, researching and writing history is a slow through-Time process – sometimes tediously so. It takes at least four years, from a standing start, to produce a big specialist, ground-breaking study of 100,000 words on a previously un-studied (or under-studied) historical topic. The exercise demands a high-level synthesis of many diverse sources, running to hundreds or even thousands. Hence the methodology is characteristically much more than a ‘reading’ of one or two key texts – although, depending upon the theme, at times a close reading of a few core documents (as in the history of political ideas) is essential too.

Mulling over meanings is an important part of the process too. History as a discipline encourages a constant thinking and rethinking, with sustained creative and intellectual input. It requires knowledge of the state of the discipline – and a close familiarity with earlier work in the chosen field of study. Best practice therefore enjoins writing, planning and revising as the project unfolds. For historical studies, ‘writing through’ is integral, rather than waiting until all the hard research graft is done and then ‘writing up’.5

The whole process is arduous and exciting, in almost equal measure. It’s constantly subject to debate and criticism from peer groups at seminars and conferences. And, crucially too, historians are invited to specify not only their own methodologies but also their own biases/assumptions/framework thoughts. This latter exercise is known as ‘self-reflexivity’. It’s often completed at the end of a project, although it’s then inserted near the start of the resultant book or essay. And that’s because writing serves to crystallise and refine (or sometimes to reject) the broad preliminary ideas, which are continually tested by the evidence.

One classic example of seriously through-Time writing comes from the classic historian Edward Gibbon. The first volume of his Decline & Fall of the Roman Empire appeared in February 1776. The sixth and final one followed in 1788. According to his autobiographical account, the gestation of his study dated from 1764. He was then sitting in the Forum at Rome, listening to Catholic monks singing vespers on Capitol Hill. The conjunction of ancient ruins and later religious commitments prompted his core theme, which controversially deplored the role of Christianity in the ending of Rome’s great empire. Hence the ‘present’ moments in which Gibbon researched, cogitated and wrote stretched over more than 20 years. When he penned the last words of the last volume, he recorded a sensation of joy. But then he was melancholic that his massive project was done.6 (Its fame and the consequent controversies last on today; and form part of the history of history).

1 For this basic point, see PJC, ‘People Sometimes Say “We Don’t Learn from the Past” – and Why that Statement is Completely Absurd’, BLOG/91 (July 2018), to which this BLOG/92 is a companion-piece.

2 See e.g. K. Jenkins, ReThinking History (1991); idem (ed.), The Postmodern History Reader (1997); C.G. Brown, Postmodernism for Historians (Harlow, 2005); A. Munslow, The Future of History (Basingstoke, 2010).

3 J. Appleby, L. Hunt and M. Jacob, Telling the Truth about History (New York, 1994); R. Evans, In Defence of History (1997); J. Tosh (ed.), Historians on History (Harlow, 2000); A. Brundage, Going to the Sources: A Guide to Historical Research and Writing (Hoboken, NJ., 2017).

4 H. Shudo, The Nanking Massacre: Fact versus Fiction – A Historian’s Quest for the Truth, transl. S. Shuppan (Tokyo, 2005); Vera Schwarcz, Bridge across Broken Time: Chinese and Jewish Cultural Memory (New Haven, 1998).

5 PJC, ‘Writing Through a Big Research Project, not Writing Up’, BLOG/60 (Dec.2015); PJC, ‘How I Write as a Historian’, BLOG/88 (April 2018).

6 R. Porter, Gibbon: Making History (1989); D.P. Womersley, Gibbon and the ‘Watchmen of the Holy City’: The Historian and his Reputation, 1776-1815 (Oxford, 2002).

For further discussion, see

To read other discussion-points, please click here

To download Monthly Blog 92 please click here

MONTHLY BLOG 91, PEOPLE SOMETIMES SAY: ‘WE DON’T LEARN FROM THE PAST’ AND WHY THAT STATEMENT IS COMPLETELY ABSURD

If citing, please kindly acknowledge copyright © Penelope J. Corfield (2018)

People sometimes say, dogmatically but absurdly: ’We don’t learn from the Past’. Oh really? So what do humans learn from, then? We don’t learn from the Future, which has yet to unfold. We do learn in and from the Present. Yet every moment of ‘Now’ constitutes an infinitesimal micro-instant an unfolding process. The Present is an unstable time-period, which is constantly morphing, nano-second by nano-second, into the Past. Humans don’t have time, in that split-second of ‘Now’, to comprehend and assimilate everything. As a result, we have, unavoidably, to learn from what has gone before: our own and others’ experiences, which are summed as everything before ‘Now’: the Past.

It’s worth reprising the status of those temporal categories. The Future, which has not yet unfolded, is not known or knowable in its entirety. That’s a definitional quality which springs from the unidirectional nature of Time. It does not mean that the Future is either entirely unknown or entirely unknowable. As an impending temporal state, it may beckon, suggest, portend. Humans are enabled to have considerable information and expectations about many significant aspects of the Future. For example, it’s clear from past experience that all living creatures will, sooner or later, die in their current corporeal form. We additionally know that tomorrow will come after today, because that is how we habitually define diurnal progression within unilinear Time. We also confidently expect that in the future two plus two will continue to equal four; and that all the corroborated laws of physics will still apply.

And we undertake calculations, based upon past data, which provide the basis for Future predictions or estimates. For example, actuarial tables, showing age-related life expectancy, indicate group probabilities, though not absolute certainties. Or, to take a different example, we know, from expert observation and calculation, that Halley’s Comet is forecast to return into sight from Earth in mid-2061. Many, though not all, people alive today will be able to tell whether that astronomical prediction turns out to be correct or not. And there’s every likelihood  that it will be.

Commemorating a successful prediction,
in the light of past experience:
a special token struck in South America in 2010 to celebrate
the predicted return to view from Planet Earth
of Halley’s Comet,
whose periodicity was first calculated by Edward Halley (1656-1742)

Yet all this (and much more) useful information about the Future is, entirely unsurprisingly, drawn from past experience, observations and calculations. As a result, humans can use the Past to illuminate and to plan for the Future, without being able to foretell it with anything like total precision.

So how about learning from the Present? It’s live, immediate, encircling, inescapably ‘real’. We all learn in our own present times – and sometimes illumination may come in a flash of understanding. One example, as Biblically recounted, is the conversion of St Paul, who in his unregenerate days was named Saul: ‘And as he journeyed, he came near Damascus; and suddenly there shined round about him a light from heaven. And he fell to the earth, and heard a voice saying unto him, “Saul, Saul, why persecutest thou me?”’1 His eyes were temporarily blinded; but spiritually he was enlightened. Before then, Saul was one of the Christians’ chief persecutors, ‘breathing out threatening and slaughter’.2 Perhaps a psychologist might suggest that his intense hostility concealed some unexpressed fascination with Christianity. Nonetheless, there was no apparent preparation, so the ‘Damascene conversion’ which turned Saul into St Paul remains the classic expression of an instant change of heart. But then he had to rethink and grow into his new role, working with those he had been attempting to expunge.

A secular case of sudden illumination appears in the fiction of Jane Austen. In Emma (1815), the protagonist, a socially confident would-be match-maker, has remained in ignorance of her own heart. She encourages her young and humble protégé, Harriet Smith, to fancy herself in love. They enjoy the prospect of romance. Then Emma suddenly learns precisely who is the object of Harriet’s affections. The result is wonderfully described.3 Emma sits in silence for several moments, in a fixed attitude, contemplating the unpleasant news:

Why was it so much worse that Harriet should be in love with Mr Knightley, than with Frank Churchill? Why was the evil so dreadfully increased by Harriet’s having some hope of a return? It darted through her, with the speed of an arrow, that Mr Knightley must marry no one but herself!

I remember first reading this novel, as a teenager, when I was as surprised as Emma at this development. Since then, I’ve reread the story many times; and I can now see the prior clues which Austen scatters through the story to alert more worldly-wise readers that George Knightley and Emma Woodhouse are a socially and personally compatible couple, acting in concert long before they both (separately) realise their true feelings. It’s a well drawn example of people learning from the past whilst ‘wising up’ in a single moment. Emma then undertakes some mortifying retrospection as she gauges her own past errors and blindness. But she is capable of learning from experience. She does; and so, rather more artlessly, does Harriet. It’s a comedy of trial-and-error as the path to wisdom.

As those examples suggest, the relationship of learning with Time is in fact a very interesting and complex one. Humans learn in their own present moments. Yet the process of learning and education as a whole has to be a through-Time endeavour. A flash of illumination needs to be mentally consolidated and ‘owned’. Otherwise it is just one of those bright ideas which can come and as quickly go.   Effective learning thus entails making oneself familiar with a subject by repetition, cogitation, debating, and lots of practice. Such through-Time application applies whether people are learning physical or intellectual skills or both. The role of perspiration, as well as inspiration, is the stuff of many mottoes: ‘practice makes perfect’; ‘if at first you don’t succeed, try and try again’; ‘stick at it’; ‘never stop learning’; ‘trudge another mile’; ‘learn from experience’.

Indeed, the entire corpus of knowledge and experience that humans have assembled over many generations is far too huge to be assimilated in an instant. (It’s actually too huge for any one individual to master. So we have to specialise and share).

So that brings the discussion back to the Past. It stretches back through Time and onwards until ‘Now’. Of course, we learn from it. Needless to say, it doesn’t follow that people always agree on messages from former times, or act wisely in the light of such information. Hence when people say: ‘We don’t learn from the Past’, they probably mean that it does not deliver one guiding message, on which everyone agrees. And that’s right. It doesn’t and there isn’t.

One further pertinent point: there are rumbling arguments around the question – is the Past alive or dead? (With a hostile implication in the sub-text that nothing can really be learned from a dead and vanished Past.) But that’s not a helpful binary. In other words, it’s a silly question. Some elements of the past have conclusively gone, while many others persist through time.4 To take just a few examples, the human genome was not invented this morning; human languages have evolved over countless generations; and the laws of physics apply throughout.

Above all, therefore, the integral meshing between Past and Present means that we, individual humans, have also come from the Past. It’s in us as well as, metaphorically speaking, behind us. Thinking of Time as running along a pathway or flowing like a river is a common human conception of temporality. Other alternatives might envisage the Past as ‘above’, ‘below’, ‘in front’, ‘behind’, or ‘nowhere specific’. The metaphor doesn’t really matter as long as we realise that it pervades everything, including ourselves.

1 Holy Bible, Acts 9: 3-4.

2 Ibid, 9:1.

3 J. Austen, Emma: A Novel (1815), ed. R. Blythe (Harmondsworth, 1969), p. 398.

4 P.J. Corfield, ‘Is the Past Dead or Alive? And the Snares of Such Binary Questions’, BLOG/62 (Feb.2016).

For further discussion, see

To read other discussion-points, please click here

To download Monthly Blog 91 please click here

MONTHLY BLOG 84, INVENTING WORDS

If citing, please kindly acknowledge copyright © Penelope J. Corfield (2017)

Speakers and writers constantly adopt and play with new words and usages, even while the deep grammatical structures of language evolve, if at all, only very slowly. I remember an English class at school when I was aged about twelve or thirteen when we were challenged to invent new words. The winning neologism was ‘puridence’. It meant: by pure coincidence. Hence, one could say ‘I walked along the pavement, puridence I slipped and fell on a banana skin’. The winner was my class-mate Audrey Turner, who has probably forgotten. (I wonder whether anyone else remembers this moment?)

2017-12 No1 slip-man-black-banana-md

Fig.1 Slip Man Black Banana:
‘Puridence I slipped and fell on a banana skin’

Another new word, invented by my partner Tony Belton on 26 October 2013, is ‘wrongaplomb’. It refers to someone who is habitually in error but always with total aplomb. It’s a great word, which immediately summons to my mind the person for whom the term was invented. But again, I expect that Tony has also forgotten. (He has). New words arrive and are shed with great ease. This is one which came and went, except for the fact that I noted it down.

No wonder that dictionary compilers find it a struggle to keep abreast. The English language, as a Germanic tongue hybridised by its conjunction with Norman French, already has a huge vocabulary, to which additions are constantly made. One optimistic proposal in the Gentleman’s Magazine in 1788 hoped to keep a check upon the process in Britain, by establishing a person or committee to devise new words for every possible contingency.1 But real-life inventions and borrowings in all living languages were (and remain) far too frequent, spontaneous and diffuse for such a system to work. The Académie française (founded 1635), which is France’s official authority on the French language, knows very well the perennial tensions between established norms and innovations.2 The ‘Immortels’, as the 40 academicians are termed, have a tricky task as they try to decide for eternity. Consequently, a prudent convention ensures that the Académie’s rulings are advisory but not binding.

For my part, I love encountering new words and guessing whether they will survive or fail. In that spirit, I have invented three of my own. The first is ‘plurilogue’. I coined this term at an academic seminar in January 2016 and then put it into a BLOG.3 It refers to multi-lateral communications across space (not so difficult in these days of easy international messaging) and through time. In particular, it evokes the way that later generations of historians constantly debate with their precursors. ‘Dialogue’ doesn’t work to explain such communications. Dead historians can’t answer back. But ‘plurilogue’ covers the multiplicity of exchanges, between living historians, and with the legacy of ideas from earlier generations.

Will the term last? I think so. Having invented it, I then decided to google (a recently-arrived verb). To my surprise, I discovered that there already is an on-line international journal of that name. It has been running since 2011. It features reviews in philosophy and political science. My initial response was to find the prior use annoying. On the other hand, that’s a selfish view. No one owns a language. Better to think that ‘plurilogue’ is a word whose time has come. Its multiple coinages are a sign of its relevance. Humans do communicate across time and space; and not just in dialogue. So ‘plurilogue’ has a tolerable chance of lasting, especially as it’s institutionalised in a journal title.

2017-12 No2 plurilogue Vol 1
A second term that I coined and published in 2007 is ‘diachromesh’.4 It defines the way that humans (and everything in the cosmos for good measure) are integrally situated in an unfolding through-Time, also known as the very long term or ‘diachronic’. That latter word is itself relatively unusual. But it has some currency among historians and archaeologists.

The ‘diachronic’ is the alternate pair to the ‘synchronic’ (the immediate fleeting moment). Hence my comment that: ‘the synchronic is always in the diachronic – in that every short-term moment contributes to a much longer term’. Equally, the conjunction operates the other way round. ‘The diachronic is always in the synchronic – in that long-term frameworks always inform the passing moment as well’.5 Therefore it follows that, just as we can refer to synchromesh gear changes, operating together in a single moment of time, so it’s relevant to think of diachromesh, effortlessly meshing each single moment into the very long-term.6

So far so good. Is diachromesh liable to last? I can’t find a journal with that name. However, the word in is circulation. Google it and see. The references are few and far between. But! For example, in an essay on the evolution of the urban high street, architectural analyst Sam Griffiths writes: ‘The spatial configuration of the grid is reticulated in space and time, a materialisation of Corfield’s (2007) “diachromesh”.’7

2017-12 No3 clock in Guildford high street

Fig.3 Guildhall Clock on Guildford High Street, marking each synchronic moment since 1683 in an urban high street, diachromeshed within its own space and time.

Lastly, I also offered the word ‘trialectics’ in 2007. Instead of cosmic history as composed of binary forces, I envisage a dynamic threefold process of continuity (persistence), gradual change (momentum) and macro-change (turbulence).8 For me, these interlocking dimensions are as integral to Time as are the standard three dimensions of Space.

Be that as it may, I was then staggered to find that the term had a pre-history, of which I was hitherto oblivious. Try web searches for trialectics in logic; ecology; and spatial theories, such as Edward Soja’s planning concept of Thirdspace.9 Again, however, it would seem that this is a word whose time has come. The fact that ‘trialectics’ is subject to a range of nuanced meanings is not a particular problem, since that happens to so many words. The core of the idea is to discard the binary of dialectics. Enough of either/or. Of point/counter-point; or thesis/antithesis. Instead, there are triple dimensions in play.

Coining new words is part of the trialectical processes that keep languages going through time. They rely upon deep continuities, whilst experiencing gradual changes – and, at the same time, facing/absorbing/rejecting the shock of the new. Luckily there is already a name for the grand outcome of this temporal mix of continuity/micro-change/macro-change. It’s called History.

1 S.I. Tucker, Protean Shape: A Study in Eighteenth-Century Vocabulary and Usage (1967), p. 104.

2 http://www.academie-francaise.fr/.

3 P.J. Corfield, ‘Does the Study of History “Progress” – and How does Plurilogue Help? BLOG/61 (Jan. 2016), www.penelopejcorfield.com/monthly-blogs/.

4 P.J. Corfield, Time and the Shape of History (2007), p. xv.

5 Ibid.

6 This assumption differs from that of a small minority of physicists and philosophers who view Time as broken, each moment sundered from the next. See e.g. J. Barbour, The End of Time: The Next Revolution in our Understanding of the Universe (1999). I might call this interpretation a case of ‘wrongaplomb’.

7 S. Griffiths, ‘The High Street as a Morphological Event’, in L. Vaughan (ed.), Suburban Urbanities: Suburbs and the Life of the High Street (2015), p. 45.

8 Corfield, Time and Shape of History, pp. 122-3, 211-16, 231, 248, 249. See also idem, ‘Time and the Historians in the Age of Relativity’, in A.C.T. Geppert and T. Kössler (eds), Obsession der Gegenwart: Zeit im 20. Jahrhundert/ Concepts of Time in the Twentieth Century (Geschichte und Gesellschaft: Sonderheft, 25, Göttingen, 2015), pp. 71-91; also available on www.penelopejcorfield.co.uk.

9 www.wikipedia.org/Edward_Soja

For further discussion, see

To read other discussion-points, please click here

To download Monthly Blog 84 please click here