WikiLeaks: five expert opinions

The New York Times has a thoughtful set of contributions from various experts on the significance of the WikiLeaks disclosures.

Evgeny Morozov, a Stanford scholar who has a book about the “dark side of Internet freedom” coming out in January, ponders the likelihood that WikiLeaks can be duplicated, and finds it unlikely.

A thousand other Web sites dedicated to leaking are unlikely to have the same effect as WikiLeaks: it would take a lot of time and effort to cultivate similar relationships with the media. Most other documents leaked to WikiLeaks do not carry the same explosive potential as candid cables written by American diplomats.

One possible future for WikiLeaks is to morph into a gigantic media intermediary — perhaps, even something of a clearing house for investigative reporting — where even low-level leaks would be matched with the appropriate journalists to pursue and report on them and, perhaps, even with appropriate N.G.O.’s to advocate on their causes. Under this model, WikiLeaks staffers would act as idea salesmen relying on one very impressive digital Rolodex.

Ron Deibert from the University of Toronto thinks that the “venomous furor” surrounding WikiLeaks, including charges of “terrorism” and calls for the assassination of Julian Assange, has to rank as “one of the biggest temper tantrums in recent years”.

Many lament the loss of individual privacy as we leave digital traces that are then harvested and collated by large organizations with ever-increasing precision. But if individuals are subject to this new ecosystem, what would make anyone think governments or organizations are immune? Blaming WikiLeaks for this state of affairs is like blaming a tremor for tectonic plate shifts.

Certainly a portion of that anger could be mitigated by the conduct of WikiLeaks itself. The cult of personality around Assange, his photoshopped image now pasted across the WikiLeaks Web site, only plays into this animosity. So do vigilante cyberattacks carried out by supporters of WikiLeaks that contribute to a climate of lawlessness and vengeance seeking. If everyone can blast Web sites and services with which they disagree into oblivion — be it WikiLeaks or MasterCard — a total information war will ensue to the detriment of the public sphere.

An organization like WikiLeaks should professionalize and depersonalize itself as much as possible. It should hold itself to the highest possible ethical standards. It should act with the utmost discretion in releasing into the public domain otherwise classified information that comes its way only on the basis of an obvious transgression of law or morality. This has not happened.

Ross Anderson, who is Professor of Security Engineering at Cambridge and the author of the standard textbook on building dependable distributed information systems, thinks that the WikiLeaks saga shows how governments never take an architectural view of security.

Your medical records should be kept in the hospital where you get treated; your bank statements should only be available in the branch you use; and while an intelligence analyst dealing with Iraq might have access to cables on Iraq, Iran and Saudi Arabia, he should have no routine access to information on Korea or Zimbabwe or Brazil. But this is in conflict with managers’ drive for ever broader control and for economies of scale.

The U.S. government has been unable to manage this trade-off, leading to regular upsets and reversals of policy. Twenty years ago, Aldrich Ames betrayed all the C.I.A.’s Russian agents; intelligence data were then carefully compartmentalized for a while. Then after 9/11, when it turned out that several of the hijackers were already known to parts of the intelligence community, data sharing was commanded. Security engineers old enough to remember Ames expected trouble, and we got it.

What’s next? Will risk aversion drive another wild swing of the pendulum, or might we get some clearer thinking about the nature and limits of power?

James Bamford, a writer and documentary producer specializing in intelligence and national security issues, thinks that the WikiLeaks disclosures are useful in forcing governments to confess.

A generation ago, government employees with Communist sympathies worried security officials. Today, after years of torture reports, black sites, Abu Ghraib, and a war founded on deception, it is the possibility that more employees might act out from a sense of moral outrage that concerns officials.

There may be more employees out there willing to leak, they fear, and how do you weed them out? Spies at least had the courtesy to keep the secrets to themselves, rather than distribute them to the world’s media giants. In a sense, WikiLeaks is forcing the U.S. government into the confessional, with the door wide open. And confession, though difficult and embarrassing, can sometimes cleanse the soul.

Fred Alford is Professor of Government at the University of Maryland and thinks that neither the Web operation WikiLeaks, nor its editor-in-chief, Julian Assange, is a whistle-blower.

Whistle-blowers are people who observe what they believe to be unethical or illegal conduct in the places where they work and report it to the media. In so doing, they put their jobs at risk.

The whistle-blower in this case is Bradley Manning, an United States Army intelligence analyst who downloaded a huge amount of government classified information, which was made public by WikiLeaks. Whether or not Manning’s act serves the greater public interest is a contentious issue, but he has been arrested and charged with unlawful disclosure of classified data.

Some have compared the role of WikiLeaks to that of The New York Times in the publication of the Pentagon Papers several decades ago. WikiLeaks is the publishing platform that leverages the vast and instantaneous distribution capacity of the Internet.

The WikiLeaks data dump challenges a long held belief by many of us who study whistle-blowing — that it is important that the whistle-blower have a name and face so that the disclosures are not considered just anonymous griping, or possibly unethical activity. The public needs to see the human face of someone who stands up and does the right thing when none of his or her colleagues dare.

But he also thinks that “for better and worse, this changes whistle-blowing as we’ve known it.”

Where the computer went

This Google video provides a company-approved tour of one of its data centres (aka server farms). I’m writing about the environmental impact of cloud computing at the moment, and rediscovered it when going through the research files for my book. It provides an interesting glimpse of the heavy engineering that lies behind cloud computing.

Why prediction is futile

At Tuesday afternoon’s Web 2.0 Summit in San Francisco, Vinod Khosla, founder of Khosla Ventures, took the stage to discuss the difference between innovation and punditry. He started off by mentioning a 1986 study that forecasted that by the year 2000, there would be just under one million cell phones. They were off by 10,000%. There were 109 milllion cell phones in the year 2000. AT&T spent $1 million and then ended up scrapping its whole cell phone business based on this forecast. Why? Because the 1980 ‘mobile phone’ was the size of a cinder block.

He then moved on to a Berkeley study that followed 80,000 forecasts over the course of 20 years and found that “experts have about the same accuracy of dart-throwing monkeys,” said Khosla. “You don’t do unreasonable things by being reasonable.”

Twitter, for example, did not exist five years ago, Khosla added. Which pundit could have predicted that a 140 character tweet would ever take off the way it did, or that a series of 140-character tweets could outline the whole culture and character of a city, like San Francisco when the Giants won the World Series?

Things like Twitter are created by innovators, not pundits.

“In every generation, you’ve seen radical shifts…Almost certainly, the next big thing won’t come from Google, Facebook, or Twitter.” To drive his home point, Khosla pointed out the fact that in the 1980s, no one thought there would be a PC in every home, and in the early 90s, no one could have predicted that email would’ve taken off. Even more shocking, the iPhone didn’t exist before 2007. “Now, it is conventional wisdom,” said Khosla.

[Source]

What’s the point of The Social Network?

This morning’s Observer column.

Lessig’s point is that it’s the open internet that should be the real hero of the story. “What’s important here,” he writes, “is that Zuckerberg’s genius could be embraced by half-a-billion people within six years of its first being launched, without – and here is the critical bit – asking the permission of anyone.” That’s true, but I think Lessig is too harsh. The message he wants the film to communicate is there in the screenplay if you look hard enough. It lies in the film’s portrayal of the contrast between what happens to unauthorised innovation on a closed, tightly controlled system and what’s possible with the open, uncontrolled architecture of the internet.

A systems view of digital preservation

The longer I’ve been around, the more concerned I become about long-term data loss — in the archival sense. What are the chances that the digital record of our current period will still be accessible in 300 years’ time? The honest answer is that we don’t know. And my guess is that it definitely won’t be available unless we take pretty rigorous steps to ensure it. Otherwise it’s posterity be damned.

It’s a big mistake to think about this as a technical problem — to regard it as a matter of bit-rot, digital media and formats. If anything, the technical aspects are the trivial aspects of the problem. The really hard questions are institutional: how can we ensure that there are organisations in place in 300 years that will be capable of taking responsibility for keeping the archive intact, safe and accessible?

Aaron Schwartz has written a really thoughtful blog post about this in which he addresses both the technical and institutional aspects. About the latter, he has this to say:

Recall that we have at least three sites in three political jurisdictions. Each site should be operated by an independent organization in that political jurisdiction. Each board should be governed by respected community members with an interest in preservation. Each board should have at least five seats and move quickly to fill any vacancies. An engineer would supervise the systems, an executive director would supervise the engineer, the board would supervise the executive director, and the public would supervise the board.

There are some basic fixed costs for operating such a system. One should calculate the high-end estimate for such costs along with high-end estimates of their growth rate and low-end estimates of the riskless interest rate and set up an endowment in that amount. The endowment would be distributed evenly to each board who would invest it in riskless securities (probably in banks whose deposits are ensured by their political systems).

Whenever someone wants to add something to the collection, you use the same procedure to figure out what to charge them, calculating the high-end cost of maintaining that much more data, and add that fee to the endowments (split evenly as before).

What would the rough cost of such a system be? Perhaps the board and other basic administrative functions would cost $100,000 a year, and the same for an executive director and an engineer. That would be $300,000 a year. Assuming a riskless real interest rate of 1%, a perpetuity for that amount would cost $30 million. Thus the cost for three such institutions would be around $100 million. Expensive, but not unmanageable. (For comparison, the Internet Archive has an annual budget of $10-15M, so this whole project could be funded until the end of time for about what 6-10 years of the Archive costs.)

Storage costs are trickier because the cost of storage and so on falls so rapidly, but a very conservative estimate would be around $2000 a gigabyte. Again, expensive but not unmanageable. For the price of a laptop, you could have a gigabyte of data preserved for perpetuity.

These are both very high-end estimates. I imagine that were someone to try operating such a system it would quickly become apparent that it could be done for much less. Indeed, I suspect a Mad Archivist could set up such a system using only hobbyist levels of money. You can recruit board members in your free time, setting up the paperwork would be a little annoying but not too expensive, and to get started you’d just need three servers. (I’ll volunteer to write the Python code.) You could then build up the endowment through the interest money left over after your lower-than-expected annual costs. (If annual interest payments ever got truly excessive, the money could go to reducing the accession costs for new material.)

Any Mad Archivists around?

Worth reading in full.

LATER: Dan Gillmor has been attending a symposium at the Library of Congress about preserving user-generated content, and has written a thoughtful piece on Salon.com about it.

The reason for libraries and archives like the Library of Congress is simple: We need a record of who we are and what we’ve said in the public sphere. We build on what we’ve learned; without understanding the past we can’t help but screw up our future.

It was easier for these archiving institutions when media consisted of a relatively small number of publications and, more recently, broadcasts. They’ve always had to make choices, but the volume of digital material is now so enormous, and expanding at a staggering rate, that it won’t be feasible, if it ever really was, for institutions like this to find, much less, collect all the relevant data.

Meanwhile, those of us creating our own media are wondering what will happen to it. We already know we can’t fully rely on technology companies to preserve our data when we create it on their sites. Just keeping backups of what we create can be difficult enough. Ensuring that it’ll remain in the public sphere — assuming we want it to remain there — is practically impossible.

Dan links to another thoughtful piece, this time by Dave Winer. Like Aaron Schwartz, Dave is concerned not just with the technological aspects of the problem, but also with the institutional side. Here are his bullet-points:

1. I want my content to be just like most of the rest of the content on the net. That way any tools create to preserve other people’s stuff will apply to mine.

2. We need long-lived organizations to take part in a system we create to allow people to future-safe their content. Examples include major universities, the US government, insurance companies. The last place we should turn is the tech industry, where entities are decidedly not long-lived. This is probably not a domain for entrepreneurship.

3. If you can afford to pay to future-safe your content, you should. An endowment is the result, which generates annuities, that keeps the archive running.

4. Rather than converting content, it would be better if it was initially created in future-safe form. That way the professor’s archive would already be preserved, from the moment he or she presses Save.

5. The format must be factored for simplicity. Our descendents are going to have to understand it. Let’s not embarass ourselves, or cause them to give up.

6. The format should probably be static HTML.

7. ??

The amateur dictator

For years, I have wanted to use speech recognition software, but have always held back because the best product – Dragon Dictate – ran only under Windows, and I am a Mac user. But the company has released a version of the program for OS X, and I’ve just installed it. And this blog post is the first I’ve ever done simply by talking to the machine.

The accuracy of Dragon Dictate is scary. The strange thing is that we used to think that successful speech recognition was a problem for Artificial Intelligence. What none of us suspected was that it was a problem that could be solved using statistics and brute-force calculation. In fact, many years ago I knew someone in Cambridge – Frank Fallside – who was one of the pioneers of this approach to speech recognition. He died tragically young, so it’s nice to see how his work eventually came to fruition.

Chatbot wears down proponents of anti-Science nonsense

Now here is an excellent use of technology.

Nigel Leck, a software developer by day, was tired of arguing with anti-science crackpots on Twitter. So, like any good programmer, he wrote a script to do it for him.

The result is the Twitter chatbot @AI_AGW. Its operation is fairly simple: Every five minutes, it searches twitter for several hundred set phrases that tend to correspond to any of the usual tired arguments about how global warming isn’t happening or humans aren’t responsible for it.

It then spits back at the twitterer who made that argument a canned response culled from a database of hundreds. The responses are matched to the argument in question — tweets about how Neptune is warming just like the earth, for example, are met with the appropriate links to scientific sources explaining why that hardly constitutes evidence that the source of global warming on earth is a warming sun.

I like this approach. It’s got lots of other applications. Now, let me see: where shall we start? There’s all that gibbering about how the bond markets will come for us if Osborne doesn’t slash public spending. And then there’s the bleating of the Irish government about how the country’s situation is “manageable”. And there’s the fantastical vapourings of the Intellectual Property lobbies…

Taking the tablets

Now it’s BlackBerry’s turn. The New York Times report explains:

The introduction of a tablet computer will not end criticism from some analysts that R.I.M. is now playing catch-up with Apple. But in a bid to distinguish the PlayBook from Apple’s iPad, Michael Lazaridis, R.I.M.’s co-chief executive, said that the new tablet contained several features requested by corporate information technology departments.

In an address to conference attendees, Mr. Lazaridis called the PlayBook “the world’s first professional tablet” and repeatedly emphasized that it was fully compatible with the special servers that corporations and governments now used to control and monitor employees’ BlackBerry devices.

While the company offered some specifics about the new device, it left many questions unanswered, most notably the tablet’s price. The company was also vague about its release date, indicating only that it would be available early next year.

Among the PlayBook’s novel features are outlets that allow it to display material on computer monitors or television sets, but Mr. Lazaridis made no effort to use them during his presentation. As animations showing the device’s features appeared above him on a giant screen, he did little more with the PlayBook in his hand than switch it on.

“It’s a very real product,” said Charles S. Golvin, principal analyst with Forrester Research. “But obviously it’s very much a work in progress.”