The takedown boom

From Ars Technica:

Piracy news site TorrentFreak reports that Google removed 75 percent more URLs in 2014 than it did the previous year. Google doesn’t tally up annual totals, but it does release weekly reports on DMCA notices, and TorrentFreak took it upon itself to add up the weekly reports. Most of the takedown requests are honored. Google has a longstanding tradition of supplying DMCA takedown notices to Chilling Effects, a website that archives such requests.

Just a few years back, the number of takedown requests could be measured in the dozens, not the millions. In 2008, Google handled 62 DMCA takedown requests, and, in that year, each request was over just one copyrighted work. In later years, DMCA notices came to ask for millions of URLs to be removed to protect multiple works.

How the network is evolving

This morning’s Observer column:

Earlier this year engineer Dr Craig Labovitz testified before the US House of Representatives judiciary subcommittee on regulatory reform, commercial and antitrust law. Labovitz is co-founder and chief executive of Deepfield, an outfit that sells software to enable companies to compile detailed analytics on traffic within their computer networks. The hearing was on the proposed merger of Comcast and Time Warner Cable and the impact it was likely to have on competition in the video and broadband market. In the landscape of dysfunctional, viciously partisan US politics, this hearing was the equivalent of rustling in the undergrowth, and yet in the course of his testimony Labovitz said something that laid bare the new realities of our networked world…

Read on…

More…

Wired had an interesting series about this shift, the first episode of which has a useful graphic illustrating the difference between most people’s mental model of the Internet, and the emerging reality.

Even if you’re not on Facebook, you are still the product

This morning’s Observer column:

The old adage “if the service is free, then you are its product” needs updating. What it signified was that web services (like Facebook, Google, Yahoo et al) that do not charge users make their money by harvesting personal and behavioural data relating to those users and selling that data to advertisers. That’s still true, of course. But a more accurate version of the adage would now read something like this: if you use the web for anything (including paying for stuff) then you are also the product, because your data is being sold on to third parties without your knowledge.

In a way, you probably already knew this. A while back you searched for, say, a digital camera on the John Lewis site. And then you noticed that wherever you went on the web after that John Lewis ads for cameras kept appearing on the site you were visiting. What you were witnessing was the output of a multibillion-dollar industry that operates below the surface of the web. Think of it as the hidden wiring of our networked world. And what it does is track you wherever you go online…

Read on

Making algorithms responsible for what they do

This morning’s Observer column:

Just over a year ago, after Edward Snowden’s revelations first hit the headlines, I participated in a debate at the Frontline Club with Sir Malcolm Rifkind, the former foreign secretary who is now MP for Kensington and Chelsea and chairman of the intelligence and security committee. Rifkind is a Scottish lawyer straight out of central casting: urbane, witty, courteous and very smart. He’s good on his feet and a master of repartee. He’s the kind of guy you would be happy to have to dinner. His only drawback is that everything he knows about information technology could be written on the back of a postage stamp in 96-point Helvetica bold…

Read on

A borderless world?

One of my mantras is that for the first 20 years of its existence (up to 1993) cyberspace was effectively a parallel universe to what John Perry Barlow called ‘Meatspace’ (aka the real world). The two universes had very little to do with one another, and were radically different in all kinds of ways. But from 1993 (when Andreessen and Bina released Mosaic, the first big web browser) onwards, the two universes began to merge, which led to the world we have now — a blended universe which has the affordances of both cyberspace and Meatspace. This is why it no longer makes sense to distinguish (as politicians still do sometimes) between the Internet and the “real world”. And it’s also why we are having so much trouble dealing with a universe in which the perils of normal life are turbocharged by the affordances of digital technology.

This morning, I came on a really interesting illustration of this. It’s about how Google Maps deal with areas of the world where there are border disputes. Turns out that there are 32 countries in the world for which Google regards the border issue as problematic. And it has adopted a typical Google approach to the problem: the borders drawn on Google’s base map of a contested area will look different depending on where in the world you happen to be viewing them from.

An example: the borders of Arunachal Pradesh, an area administered by India but claimed as a part of Tibet by China. The region is shown as part of India when viewed from an Indian IP address, as part of China when viewed from China, and as distinct from both countries when viewed from the US.

There’s a nice animation in the piece. Worth checking out.

Yay! Gmail to get end-to-end encryption

This has been a long time coming — properly encrypted Gmail — but it’s very welcome. Here’s the relevant extract from the Google security blog:

Today, we’re adding to that list the alpha version of a new tool. It’s called End-to-End and it’s a Chrome extension intended for users who need additional security beyond what we already provide.

“End-to-end” encryption means data leaving your browser will be encrypted until the message’s intended recipient decrypts it, and that similarly encrypted messages sent to you will remain that way until you decrypt them in your browser.

While end-to-end encryption tools like PGP and GnuPG have been around for a long time, they require a great deal of technical know-how and manual effort to use. To help make this kind of encryption a bit easier, we’re releasing code for a new Chrome extension that uses OpenPGP, an open standard supported by many existing encryption tools.

However, you won’t find the End-to-End extension in the Chrome Web Store quite yet; we’re just sharing the code today so that the community can test and evaluate it, helping us make sure that it’s as secure as it needs to be before people start relying on it. (And we mean it: our Vulnerability Reward Program offers financial awards for finding security bugs in Google code, including End-to-End.)

Once we feel that the extension is ready for primetime, we’ll make it available in the Chrome Web Store, and anyone will be able to use it to send and receive end-to-end encrypted emails through their existing web-based email provider.

We recognize that this sort of encryption will probably only be used for very sensitive messages or by those who need added protection. But we hope that the End-to-End extension will make it quicker and easier for people to get that extra layer of security should they need it.

Cars as services, not possessions?

This morning’s Observer column.

We now know that the implications of the driverless cars’ safety record were not lost on Google either. Last week the company rolled out its latest variation on the autonomous vehicle theme. This is a two-seater, pod-like vehicle which scoots around on small wheels. It looks, in fact, like something out of the Enid Blyton Noddy stories. The promotional video shows a cheery group of baby-boomers summoning these mobile pods using smartphones. The pods whizz up obligingly and stop politely, waiting to be boarded. The folks get in, fasten their seatbelts and look around for steering wheel, gear shift, brake pedals etc.

And then we come to the punchline: none of these things exist on the pod! Instead there are two buttons, one marked “Start” and the other marked “Stop”. There is also a horizontal computer screen which doubtless enables these brave new motorists to conduct Google searches while on the move. The implications are starkly clear: Google has decided that the safest things to do is to eliminate the human driver altogether.

At this point it would be only, er, human to bristle at the temerity of these geeks. Who do they think they are?

Read on

Google privacy ruling: the thin end of a censorship wedge?

This morning’s Observer column.

Sooner or later, every argument about regulation of the internet comes down to the same question: is this the thin end of the wedge or not? We saw a dramatic illustration last week when the European court of justice handed down a judgment on a case involving a Spanish lawyer, one Mario Costeja González, who objected that entering his name in Google’s search engine brought up embarrassing information about his past (that one of his properties had been the subject of a repossession)…

Read on

LATER

Three interesting — and usefully diverse — angles on the ECJ decision.

  • Daithi Mac Sitigh points out that the decision highlights the tensions between EU and US law. “This is particularly significant”, he says, “given that most of the major global players in social networking and e-commerce operate out of the US but also do a huge amount of business in Europe.”

Google’s first line of defence was that its activities were not subject to the Data Protection Directive. It argued that its search engine was not a business carried out within the European Union. Google Spain was clearly subject to EU law, but Google argued that it sells advertising rather than running a search engine.

The court was asked to consider whether Google might be subject to the Directive under various circumstances. A possible link was the use of equipment in the EU, through gathering information from EU-based web servers or using relevant domain names (such as google.es). Another suggestion was that a case should be brought at its “centre of gravity”, taking into account where the people making the requests to delete data have their interests.

But the court never reached these points. Instead, it found the overseas-based search engine and the Spain-based seller of advertising were “inextricably linked”. As such, Google was found to be established in Spain and subject to the directive.

The message being sent was an important one. Although this ruling is specific to the field of data protection, it suggests that if you want to do business in the EU, a corporate structure that purports to shield your activities from EU law will not necessarily protect you from having to comply with local legislation. This may explain the panicked tone of some of the reaction to the decision.

  • In an extraordinary piece, “Right to Forget a Genocide”, Zeynep Tufekci muses about how (Belgian) colonial imposition of ID cards on Rwandan citizens was instrumental in facilitating genocide.

It may seem like an extreme jump, from drunken adolescent photos to genocide and ethnic cleansing, but the shape, and filters, of a society’s memory is always more than just about individual embarrassment or advancement. What we know about people, and how easily we can identify or classify them, is consequential far beyond jobs and dates, and in some contexts may make the difference between life and death.

“Practical obscurity”—the legal term for information that was available, but not easily—has died in most rich countries within just about a decade. Court records and criminal histories, which were only accessible to the highly-motivated, are now there at the click of a mouse. Further, what is “less obscure” has greatly expanded: using our online data, algorithms can identify information about a person, such as sexual orientation and political affiliation, even if that person never disclosed them.

In that context, take Rwanda, a country many think about in conjunction with the horrific genocide 20 years ago during which more than 800,000 people were killed—in just about one hundred days. Often, stories of ethnic cleansing and genocide get told in a context of “ancient hatreds,” but the truth of it is often much uglier, and much less ancient. It was the brutal colonizer of Rwanda, Belgium, that imposed strict ethnicity-based divisions in a place where identity tended to be more fluid and mixed. Worse, it imposed a national ID system that identified each person as belonging to Hutu, Tutsi or Twa and forever freezing them in that place. [For a detailed history of the construction of identity in Rwanda read this book, and for the conduct of colonial Belgium, Rwanda’s colonizer, read this one.]

Few years before the genocide, some NGOs had urged that Rwanda “forget” ethnicity, erasing them from ID cards.

They were not listened to.

During the genocide, it was those ID cards that were asked for at each checkpoint, and it was those ID cards that identified the Tutsis, most of whom were slaughtered on the spot. The ID cards closed off any avenue of “passing” a checkpoint. Ethnicity, a concept that did not at all fit neatly into the region’s complex identity configuration, became the deadly division that underlined one of the 20th century’s worst moments. The ID cards doomed and fueled the combustion of mass murder.

  • Finally, there’s a piece in Wired by Julia Powles arguing that “The immediate reaction to the decision has been, on the whole, negative. At best, it is reckoned to be hopelessly unworkable. At worst, critics pan it as censorship. While there is much to deplore, I would argue that there are some important things we can gain from this decision before casting it roughly aside.”

What this case should ideally provoke is an unflinching reflection on our contemporary digital reality of walled gardens, commercial truth engines, and silent stewards of censorship. The CJEU is painfully aware of the impact of search engines (and ‘The’ search engine, in particular). But we as a society should think about the hard sociopolitical problems that they pose. Search engines are catalogues, or maps, of human knowledge, sentiments, joys, sorrows, and venom. Silently, with economic drivers and unofficial sanction, they shape our lives and our interactions.

The fact of the matter here is that if there is anyone that is up to the challenge of respecting this ruling creatively, Google is. But if early indications are anything to go by, there’s a danger that we’ll unwittingly save Google from having to do so, either through rejecting the decision in practical or legal terms; through allowing Google to retreat “within the framework of their responsibilities, powers and capabilities” (which could have other unwanted effects and unchecked power, by contrast with transparent legal mechanisms); or through working the “right to be forgotten” out of law through the revised Data Protection Regulation, all under the appealing but ultimately misguided banner of preventing censorship.

There is, Powles argues, a possible technical fix for this — implementation of a ‘right to reply’ in search engine results.

An all-round better solution than “forgetting”, “erasure”, or “take-down”, with all of the attendant issues with free speech and the rights of other internet users, is a “right to reply” within the notion of “rectification”. This would be a tech-enabled solution: a capacity to associate metadata, perhaps in the form of another link, to any data that is inaccurate, out of date, or incomplete, so that the individual concerned can tell the “other side” of the story.

We have the technology to implement such solutions right now. In fact, we’ve done a mock-up envisaging how such an approach could be implemented.

Search results could be tagged to indicate that a reply has been lodged, much as we see with sponsored content on social media platforms. Something like this, for example:

Forgotten

(Thanks to Charles Arthur for the Tufekci and Powles links.)

What happens when algorithms decide what should be passed on?

One of the things we’re interested in on our research project is how rumours, news, information (and mis-information) can spread with astonishing speed across the world as a result of the Internet. Up to now I had been mostly working on the assumption that the fundamental mechanism involved is always something like the ‘retweet’ in Twitter — i.e. people coming on something that they wanted to pass on to others for whatever reason. So human agency was the key factor in viral retransmission of memes.

But I’ve just seen an interesting article in the Boston Globe which suggests that we need to think of the ‘retweeting’ effect in wider terms.

A surprise awaited Facebook users who recently clicked on a link to read a story about Michelle Obama’s encounter with a 10-year-old girl whose father was jobless.

Facebook responded to the click by offering what it called “related articles.” These included one that alleged a Secret Service officer had found the president and his wife having “S*X in Oval Office,” and another that said “Barack has lost all control of Michelle” and was considering divorce.

A Facebook spokeswoman did not try to defend the content, much of which was clearly false, but instead said there was a simple explanation for why such stories are pushed on readers. In a word: algorithms.

The stories, in other words, apparently are selected by Facebook based on mathematical calculations that rely on word association and the popularity of an article. No effort is made to vet or verify the content.

This prompted a comment from my former Observer colleague, Emily Bell, who now runs the Tow Center at Columbia. “They have really screwed up,” she told the Globe. “If you are spreading false information, you have a serious problem on your hands. They shouldn’t be recommending stories until they have got it figured out.”

She’s right, of course. A world in which algorithms decided what was ‘newsworthy’ would be a very strange place. But we might get find ourselves living in such a world, because Facebook won’t take responsibility for its algorithms, any more than Google will take responsibility for YouTube videos. These companies want the status of common carriers because otherwise they have to assume legal responsibility for the messages they circulate. And having to check everything that goes through their servers is simply not feasible.