In praise of essays

One of the interesting side-effects of Twitter is that it liberates bloggers from one of their original duties — that of providing a set of links to interesting stuff elsewhere on the Web. That, after all, is where the genre originated — as weblogs which were just that, a list of links to other sites. In that sense, the first blogger was Tim Berners-Lee, the Web’s inventor, because one of the first pages he published on CERN’s server (info.cern.ch) was a daily-updated list of other web servers.

As the web grew, and pioneers like Dave Winer began to explore what could be done with this new publishing medium, a consensus emerged that a weblog should be more than just a set of links — that there should be at least some kind of annotation or comment accompanying the links. And so Blogging was born and rapidly morphed into a literary form which uneasily mixed link-logs with heavily hyperlinked prose pieces of varying length, profundity and authority.

I started blogging in the mid-1990s because I needed a way of creating a hyperlinked notebook to keep track of my online reading and to support my academic and journalistic work. I had discovered that if one spent any amount of time on the Web then simple, browser-based bookmarking rapidly became dysfunctional. I first used a tool called AOLPress which was free and lightweight (and which, I learned much later from TB-L’s memoir, had very respectable antecedents). For the first few years, my blog was effectively private — it was hosted on my own computers and not available on the Web. This made sense at the time because it really was more like a lab notebook than a literary product: it was a way of enabling me to escape the consequences of a poor memory, especially once I’d put a search engine onto it. I knew from then on that if I’d written about it on my blog then I would always be able to retrieve it later.

Memex went public towards the end of the 1990s after I’d started to use Dave Winer’s Userland software. But going public involved an uneasy compromise. The terseness of the private blog had to be softened, somehow, by some degree of elaboration, explanation or exposition. Otherwise, readers might have no idea of what lay behind a particular link or observation. So from the moment Memex went public it’s oscillated between weblog and blog, with a strong bias towards the former. Every so often, I would post extended pieces which were more or less polished (usually the latter), but for most of the time Memex has been mainly a rushed, idiosyncratic guide to things I have found interesting, instructive or significant. In that sense it sits at the opposite end of the spectrum from blogs whose posts are well-formed, carefully-crafted mini-essays: like, say, those written by the venerable Chicago firm of Becker & Posner, or by Paul Graham, Ed Felten, Clay Shirky, Diane Coyle, Martin Weller, Bill Thompson, or Sean French & Nicci Gerrard — to pick just a few names at random from my blogroll.

The great thing about social bookmarking services like del.icio.us and Twitter is that they could liberate bloggers from the weblogging side of their lives and allow them to concentrate on, well, online essays. Which of course raises the question of what is an ‘essay’. As chance would have it, this is a subject discussed by Zadie Smith in the introduction to her splendid new collection of the things, Changing My Mind: Occasional Essays.

As a noun, Smith observes, the word ‘essay’ has had “an unstable history, shape-shifting over the centuries in its little corner of the OED”.

For Samuel Johnson in 1755 it is: “A loose sally of the mind; an irregular undigested piece; not a regularly and orderly composition.” And if this looks to us like one of Johnson’s lexical eccentricities, we’re chastened to find Joseph Addison, of all people, in agreement (“The wildness of these compositions that go by the name of essays”) and behind them both three centuries of vaguely negative connotation. Beginning in the 1500s an essay is: the action or process of trying or testing; a sample, an example; a rehearsal; an attempt or endeavour; a trying to do something; a rough copy; a first draft. Not until the mid 19th century does it take on its familiar, neutral ring: “a composition more or less elaborate in style, though limited in range.”

Johnson’s definition seems to me to fit many of the longer posts one finds on the best blogs. They’re full of ideas, but not quite polished or honed: work in progress rather than finished products. As someone who has written a weekly mini-essay for (print) publication in a national magazine or newspaper for 50 weeks a year since 1982, I’m often struck by the differences between my Observer column and what I post here. In part, this is a response to the constraints laid upon one by a print publication; although my column is published (and widely read) online, its most important parameter — length — is set by the requirements of the page on which it sits in the print edition of the newspaper. So when I sit down to write a column, I know that whatever I have to say has to be fitted into 800 words.

Such a limitation makes no sense on the Web, where space is, in theory, infinite. But in fact the discipline imposed by the 800-word limit seems to me to be a beneficial thing. For one thing, it discourages prolixity and encourages brevity. As someone who agrees with Wittgenstein’s dictum that “if a thing can be said then it can be said simply clearly”* I rather like that.

An 800-word limit has other benefits too. It reminds one, for example, that readers’ time is precious and that one shouldn’t waste it. Space may be abundant on the Web, but attention is an increasingly scarce resource in this networked world. As Herbert Simon put it:

In an information-rich world, the wealth of information means a dearth of something else: a scarcity of whatever it is that information consumes, What information consumes is rather obvious: it consumes the attention of its recipients. Hence a wealth of information creates a poverty of attention and a need to allocate that attention efficiently among the overabundance of information sources that might consume it.

Simon wrote that in 1971, long before the Web was created and just when the Arpanet was getting into its stride, but it seems spot-on for today’s circumstances.

The other difference between writing for print and writing for one’s blog is that there comes a moment with the print essay when it has to be ‘finished’ and dispatched to the sub-editors: there’s an ‘end-point’, in other words. But, in a sense, a blog post is never ‘finished’; there’s always the possible of ongoing revision in the light of comments, or second thoughts, or sheer, unreasoning loss of nerve. You could say, therefore, that writing for print is like sculpting in stone, whereas writing for a blog is like sculpting in jelly that hasn’t quite set.

So kindly stand aside while I pour…

*FOOTNOTE: Shortly after this was posted, my learned colleague, Doug Clow, corrected the Wittgenstein quote.

Majority Of Republicans Think Obama Didn’t Actually Win 2008 Election

Wow! There are more nutters in the US than even I had supposed. According to this report,

The new national poll from Public Policy Polling (D) has an astonishing number about paranoia among the GOP base: Republicans do not think President Obama actually won the 2008 election — instead, ACORN stole it.

This number goes a long way towards explaining the anger of the Tea Party crowd. They not only think Obama’s agenda is against America, but they don’t think he was actually the choice of the American people at all! Interestingly, NY-23 Conservative candidate Doug Hoffman is now accusing ACORN of stealing his race, and Fox News personalities have often speculated about ACORN stealing the 2008 Minnesota Senate race for Al Franken.

The poll asked this question: “Do you think that Barack Obama legitimately won the Presidential election last year, or do you think that ACORN stole it for him?” The overall top-line is legitimately won 62%, ACORN stole it 26%.

Among Republicans, however, only 27% say Obama actually won the race, with 52% — an outright majority — saying that ACORN stole it, and 21% are undecided. Among McCain voters, the breakdown is 31%-49%-20%. By comparison, independents weigh in at 72%-18%-10%, and Democrats are 86%-9%-4%.

Now, the obvious comparison would be that many Democrats felt that George W. Bush didn’t legitimately win the 2000 election. But there are some clear differences.

First of all, Al Gore empirically won the national popular vote in 2000, and lost in a disputed recount process in Florida. By comparison, John McCain lost the national popular vote by a 53%-46% margin.

In order to believe that Obama wasn’t the true winner of the 2008 election, one would have to think that ACORN (and perhaps other groups) stuffed ballots to the tune of over 9.5 million votes, Obama’s national margin.

What’s 26 per cent of 200 million? And this is a country with nuclear weapons.

Apple’s Mistake

Paul Graham is a terrific, perceptive essayist. (If you haven’t read his collection Hackers & Painters: Big Ideas from the Computer Age: Essays on the Art of Programming then might I humbly suggest a visit to Amazon?)

His latest essay on how Apple is treating the programmers who develop Apps for the iPhone/iTouch is characteristically acute.

This is how it opens:

I don’t think Apple realizes how badly the App Store approval process is broken. Or rather, I don’t think they realize how much it matters that it’s broken.

The way Apple runs the App Store has harmed their reputation with programmers more than anything else they’ve ever done. Their reputation with programmers used to be great. It used to be the most common complaint you heard about Apple was that their fans admired them too uncritically. The App Store has changed that. Now a lot of programmers have started to see Apple as evil.

How much of the goodwill Apple once had with programmers have they lost over the App Store? A third? Half? And that’s just so far. The App Store is an ongoing karma leak…

Take a break. Grab a coffee. And read the whole piece.

The only way to break Apple’s stranglehold on the Apps business is to find a way of making the Android platform attractive to developers. The problem is — as Graham points out — that most geeks have iPhones and we know from long experience that the best software comes from programmers “scratching an itch” (as Eric Raymond put it). So maybe an intelligent strategy would be for Google (or Motorola or other handset manufacturers who aim to offer Android phones) to identify developers and offer them free Android phones.

Google’s new Operating System explained

Very neat exposition. Just the right number of half-truths.

Google’s now saying that Chrome will be appearing on netbooks in the latter half of 2010. It’ll make waves.

UPDATE: Technology Review has a useful outline of the OS. It contains this interesting snippet:

The operating system won’t be available for download, however. Because of its tight integration between software and hardware, users will have to buy a Chrome device from one of Google’s partners in order to use it. Google plans to give partners strict hardware requirements for the devices, specifying particular wireless cards and other components.

The code is, however, open source and anyone can run it on a virtual machine.

The New York Times take on it is that this is a direct challenge to Microsoft:

“Hundreds of millions of users are living on the cloud,” said Sundar Pichai, a vice president for product management at Google in charge of Chrome. Every program that users enjoy on their PCs today, Mr. Pichai said, will soon be available as a Web application. “The trend is very, very clear,” he said.

While Microsoft and others say they believe that cloud-based programs will coexist with traditional PC software, Google has often said that Web applications will replace all desktop software, another area that Microsoft dominates. Machines running the Chrome operating system, which initially will be limited to lightweight, portable computers known as netbooks, will not run any desktop applications other than the Chrome browser.

But even Mr. Pichai said that devices on the Chrome operating system were likely to be used, at least at first, as a complement to users’ more powerful computers at home.

Analysts said that the Chrome operating system could pose a challenge to Microsoft over the long term but said that Microsoft was not sitting still.

In truth, this isn’t a paradigm shift. That happened ages ago with things like Hotmail and Gmail and Google Docs as the network supplanted the PC as the computer. The Chrome OS is just the first major operating system product built on the shift.

Cameronspeak

This is a wordcloud (generated using tagcrowd) from the text of a major speech made by the Tory Leader at the Open University on May 26 this year. Might be an interesting tool for tracking changes in rhetoric — and perhaps even policy.

Remembering Peter Drucker

I’ve always thought that Peter Drucker is the only writer one could legitimately call a “management guru” (though Charles Handy runs him close). So it’s nice to come on this essay in the current edition of the Economist. Excerpt:

The world’s great business schools have replaced Oxbridge as the nurseries of the global elite. The management-consulting industry will earn revenues of $300 billion this year. Management books regularly top the bestseller lists. Management gurus can command $60,000 a speech.

Yet the practitioners of this great industry continue to suffer from a severe case of status anxiety. This is partly because the management business has always been prey to fads and fraudsters. But it is also because the respectable end of the business seems to lack what Yorkshire folk call “bottom”. Consultants and business-school professors are forever discovering great ideas, like re-engineering, that turn to dust, and wonderful companies, like Enron, that burst into flames.

Peter Drucker is the perfect antidote to such anxiety. He was a genuine intellectual who, during his early years, rubbed shoulders with the likes of Ludwig Wittgenstein, John Maynard Keynes and Joseph Schumpeter. He illustrated his arguments with examples from medieval history or 18th-century English literature. He remained at the top of his game for more than 60 years, advising generations of bosses and avoiding being ensnared by fashion. He constantly tried to relate the day-to-day challenges of business to huge social and economic trends such as the rise of “knowledge workers” and the resurgence of Asia.

But Drucker was more than just an antidote to status anxiety. He was also an apostle for management. He argued that management is one of the most important engines of human progress: “the organ that converts a mob into an organisation and human effort into performance”.