This post is adapted from a talk I gave to the annual conference of the Australasian Association of Digital Humanities in Hobart on 21st June 2016. It was a great conference with some great papers, leading to some great discussions. This particular talk generated perhaps more heat than light, but the main point seems important to me.
The
Digital Humanities is a funny beast. I tend to think of it as something
of a pantomime horse – with criticism, distant reading and literary theory
occupying the front end – all neighing and foot stamping at the MLA each year
(and in the Los Angeles Review of Books) - while history, geography and
library science are stuck in the rear – doing the hard work of creating new
digital resources, and testing new tools. Firmly at the back end of this
arrangement, I spend much of my time hoping that the angry debates in the
front don’t result in too many ructions behind.
But as a
result of this weird portmanteau existence the Digital Humanities – its debates
and its aspirations - has been largely about text. As Matthew Kirchenbaum
has noticed – much of it can be found in the English department. Its
origins are always located in the work of Father Busa and its greatest stars
from Franco Moretti onwards, keep us focussed on the ‘distant reading’ of words.
Indeed, the object of study for most Digital Humanists remains the inherited
text of Western culture – now available for recalculation via Google Books,
ECCO, EEBO, and Project Gutenberg.
And
because Digital Humanities is being led from areas of the academy that take as
its object of study a canonical set of texts (however extensive and contested);
we have been naturally led to use tools that privilege text analysis and to
ignore methodologies that are focussed elsewhere. The popular tools on
the block are topic modelling of text, and network analysis based on the
natural language processing, of text. This is particularly
true in North America where subjects such as geography do not have as strong an
institutional presence as in Europe and much of the rest of the world, and
where the spatial and sonic turns in the humanities feel less well established.
This emphasis on text tends to make the
Digital Humanities feel rather safer than it should. While the digital humanities is
frequently cited for its disruptive potential - its ‘affordances’ – it is
inherently conservative about what constitutes a legitimate subject, and has
breathed new life into areas that forty years ago lready felt moribund. The Enlightenment, and the papers of Newton,
Austen, Bentham and Darwin – all dead writers of an elite stamp - have been
revived and their ‘texts’ been made hyper-available.
In part, this is just about the rhythms of the
academy. The Digital Humanities arose just as post-modernism and second wave
feminist criticism seemed to exit stage left, and with them, much of the
imperative to critique the canon. But it
is also a result of underlying economic structures and the technologies of
twentieth-century librarianship. We very
seldom acknowledge it, but the direction of our work is frequently determined
and universally facilitated by the for-profit commercial information sector – by
the likes of ProQuest, Google and Elsevier, Ancestry.com and Cengage Gale. And they in turn are the product of a
hundred-year history that has shaped what is available to all researchers in the humanities.
If you want to know why, for example, The Times Digital Archive was the first
major newspaper available online; if you want to know why early modern English
books came next; if you want to know why Indian and African and South American literature is not available in the same way, it is down to
selections made by these companies, and selections made, not last year, but a
hundred years ago.
The current
digital landscape is actually a reflection of an older underlying project,
and an older technology. Perhaps the
biggest influence on what is available to researchers online – the biggest selection
bias involved - is just a ghost of commercially produced, for profit,
microfilm. In other words, we have text
because that is what people thought was important in 1906 or 1927, or 1935. We tend to forget that microfilm was the great new technology of the twentieth century - and was itself part of an apparently radical disruptive intellectual project. It is worthwhile remembering the details.
In 1906 Paul Otlet and Robert
Goldschmidt proposed the livre microphotographique – library of
microfilm – as a World Center Library of Juridical, Social and Cultural
Documentation. This was to be the
ultimate universal library and knowledge machine – the library at Alexandria
made new – and it was made possible by microfilm.
Later perfected for commercial uses, by the late 1920s, microfilm was the methodology of choice used by the Library of Congress to film and republish some 3 million volumes from the British Library between 1927 and 1935; and in 1935, Kodak started filming The Times on a commercial basis.
Later perfected for commercial uses, by the late 1920s, microfilm was the methodology of choice used by the Library of Congress to film and republish some 3 million volumes from the British Library between 1927 and 1935; and in 1935, Kodak started filming The Times on a commercial basis.
And this pre-history of heritage material on the web is relevant for a simple reason. It costs less than a penny per page to generate a digital image from a microfilm. It is automated to the point that all you do is feed the reel into a machine wait. By way of comparison, it costs around 15 pence per page to generate a similar image from a real book – even with modern automation, and three times that again capture a page of manuscript in an archive. For many of the projects designed during the first decade of the web, it was cheaper to have material microfilmed first, as a first step in digitisation.
Seventeenth and eighteenth century books are available online precisely because the Library of Congress microfilmed them in in the 1920s; and The Times Digital Archive is available because Kodak microfilmed it over eighty years ago. Chinese and Arabic literature is not available in the same way because the Library of Congress and Kodak and their ilk decided it was not important. Pro-Quest, the multi-billion pound corporation that supplies half the material used by Digital Humanities scholars, started as University Microfilms International in 1938.
In other
words, what happened in the twentieth century – the aspiration to create a
particular kind of universal library, and to commercialise world culture (and to
a 1930s mind, this meant male and European culture) – essentially shapes what
is now available on line. This is why most of the material we currently
have is in black and white instead of colour. And most importantly, it is why we have text; and in particular,
canonical texts in English.
And if
the Digital Humanities was really only the front end of that pantomime horse,
this would not be that big a deal. But the Digital Humanities is also
the back end – all the people creating the infrastructure that defines world
culture online. If you ask an undergraduate (or most humanities
professors) about their research practises, it rapidly becomes clear that hard
copy wood pulp has been replaced by digital materials. What we study is
what we can find on line.
In part,
the selection bias driven by the role of microfilm and the textual bias this
implies, just means that like the humanities in general, the digital sort is
inherently, and institutionally, Western centric, elitist and racist.
Rich white people produced the text that the humanities tend to study and despite the heroic multi-generation effort that has sought to recover
female voices; or projects seeking to give new voice to the poor – from below -
this selective intellectual landscape remains.
In other
words, the textual Digital Humanities offers a superficial and faux radicalism
that effectively re-enforces the conservative character of much humanities
research. The Digital Humanities' problem in
recruiting beyond white and privileged practitioners is not just down to the
boorish cultures of code – rude male children being unwelcoming - but a
result of its object of study.
All of
which is just by way of introducing the real subject of this post - that for us
to actually grasp the ‘affordances’ that the digital makes possible we really
need to change that ‘object of study’ and move beyond microfilmed
cultures.
And that when we add space and place, time and sound, to our analysis, and when we start from a hundred other places than the English department – from geography and archaeology, to quantitative biology and informatics – we can create something that is more compelling, more revealing and more powerful – and arguably more inclusive and democratic along the way.
And that when we add space and place, time and sound, to our analysis, and when we start from a hundred other places than the English department – from geography and archaeology, to quantitative biology and informatics – we can create something that is more compelling, more revealing and more powerful – and arguably more inclusive and democratic along the way.
By way of pursuing this idea, I want to go through a few of the different ways
new tools and approaches create real opportunities to move beyond the analysis
of ‘text’ to something more ambitious; and in the process attack that very real
inherent bias – and inherent conservatism - that the ‘textual’ humanities
brings with it.
The rain
falls on every head – and I
just want explore how we can move beyond the elite and the Western, the
privileged and the male.
In
the humanities we think of digitisation of text,but in a dozen other fields, they
are digitising different components of the physical world. And when
everything is digital – when all forms of stuff come to us down a single
pipeline - everything can be inter-related in new ways. The web and
the internet simply provides a context
in which image, sound, video and text are brought onto a single page.
Consider for a moment the ‘Haptic Cow’ project from the Royal
Veterinary College in London. In this instance they have developed a full
scale ‘haptic’ representation of a cow in labour, facing a difficult birth,
which allows students to physically engage and experience the process of
manipulating a calf in situ. Imagine this technology applied to a more
historical event, or process, or experience. It suggests that the object
of study can be different, and should include the haptic - the feel and heft of
a thing in your hand. This is being coded for millions of objects through
3d scanning; but we do not yet have an effective way of incorporating that 3d
encoding into our reading of the past.
And if
we can ‘feel’ an object, it changes how we read the text that comes with it; or
the experience that text encodes. The world would look and feel very
different if we organised it around those objects – the inherited texts
attached to them perhaps - but those objects’ origin and materiality forming
the core of the meaning we seek to interrogate. We can use the technology
to think harder about the changing nature of work, or punishment, the ‘feel’ of
oppression and luxury. Museums and
collections - the catacombs of culture - are undoubtedly just as powerfully selective and controlling as the
unseen hand of the publishers and archivists; but in stepping beyond text, we
can hope to play the museums off against the text.
The same
could be said of the aural - that weird world of sound on which we continually
impose the order of language, music and meaning; but which is in fact a stream
of sensations filtered through place and culture. For people working in
musicology there feels to be a ‘sonic turn’ in the humanities, but most of us
have paid it little heed.
There
are projects like the Virtual St Paul's Cross, which allows
you to ‘hear’ John Donne’s sermons from the 1620s, from different vantage
points around the yard. Donne is a dead white man par excellence, but
the project changes how we imagine the text and the event. And again begins to navigate
that normally unbridgeable space between text and the material world to help give us access to the experience of the beggar in the
crowd, of women, children and the historically unvoiced.
For myself,
I want to understand a sermon heard in the precise church in which it was
delivered; a political speech in the field, or parliamentary chamber; or an
impassioned defence in the squalid courtroom in which it was enacted, or under
an African judgement tree – with the weather and the smell thrown in. And
I want to hear it from the back of the hall, through the ears of a child or a
servant.
This would help challenge us to think harder and
differently about text that purport to represent speech, and text that sits
between the mind and the page. Recorded voice – even in the form of text
– is inherently more quotidian, is inherently more likely to give us access
to the 90 percent of the population whose voices are recorded, but whose 'text' is not. Text
recording speech is different to text produced by the elite power users of the
technology of writing – who write directly from mind to page. This at
least shifts us a bit – from text, to voice.
Similarly,
in the work of people such as Ian Gregory, we can see the beginnings of new
ways of reading both the landscape, and the textual leavings of the dead in the
landscape. His projects on mapping the Lakeland poets; and
mapping 19th century government reports, imply a new and different
kind of reading.
What
happens to a traveller's journal when it is mapped onto a landscape? What
happens to a landscape painting when we can see both its reference landscape,
and the studio in which it was completed? What happens to even text, when
it is understood to encode a basic geographical relationship? How do we
understand a conversation on a walk when we can map its phrases and exchanges
against the earth’s surface? And what forms of analysis can undertake
when each journey, each neighbourhood, each street and room, are available to
add to the text associated with them?
The rain
falls on every head.
All of
which is to state the obvious. There are lots of new technologies that
change how we connect with historical evidence – whether that is text or
something more interesting; and that we increasingly access it all via that
single remarkable pipeline that is the online and the digital.
But it
strikes me that adding these new dimensions to the object of study allows us
to do something important. I have spent the last thirty-eight years
working on a ‘history from below’ focused on the lives of eighteenth century
London’s working people. And what I want to suggest is that these new
dimensions and methodologies actually make that project fundamentally more
possible; and by extension makes the larger project of recovering the voices
and experience of the voiceless dead, more possible. When you add in the
haptic, the mapped and the geographical, the aural and the 3D, what you
actually end up with is a world in which non-elite – and non-western - people
are newly available in a new way. You also move from a kind of history as
explanation, to history as empathy – across cultures and genders, across time
and space.
Sound
and space and place, are fundamentally more intellectually democratic than text.
90% our inherited canon is inherited from rich dead white men; and yet the
thronging multitude who stood in St Paul’s Churchyard; the quotidian hoards who
walked through the streets and listened to the ballad singers, experienced
something that we can now recover. The sound of judgement as experienced
by the women and men who stood trial at the Old Bailey, and their voice of
defiance, can be recovered. And even the cold and wind of a weather that
can now be captured day by day for a quarter of a millennium; can be added to
the democratic possibilities new digital resources allow. Add in the
objects in the museums, the sounds of the ships, and their course through the
oceans; the measurable experience of labour, and imprisonment, the joy of music
and movement, the inherited landscapes, bearing all the marks of the toil of
the voiceless dead, and you end up with something new. The material world – in
digital – gives us access to the rest of the world, and begins to create tools
that speak to the 99% of the world’s population who, in 1700 or 1800 did not
read or write, and did not leave easy traces for us to follow. The Digital Humanities in Three Dimensions, challenges us and empowers us, to write a different, more inclusive, kind of history.
The rain
falls on every head.