What We Can Learn From the Guardian’s New Open Platform

British national paper The Guardian isn’t the kind of tech-savvy enterprise one would normally look to for guidance on digital issues or Internet-related topics. For one thing, it’s not a startup — it’s a 190-year-old newspaper. And it’s not based in Palo Alto or SoMa, but in London Manchester, England. The newspaper company, however, is doing something fairly revolutionary. In a nutshell, The Guardian has completely rethought the fundamental nature of its business — something it has effectively been forced to do, as many media entities have, by the nature of the Internet — and, as a result, has altered the way it thinks about value creation and where that comes from.

Enter The Guardian’s “Open Platform,” which launched last week and involves an open API (application programming interface) that developers can use to integrate Guardian content into services and applications. The newspaper company has been running a beta version of the platform for a little over a year now, but took the experimental label off the project on Thursday and announced that it is “open for business.” By that The Guardian means it is looking for partners who want to use its content in return for either licensing fees or a revenue-sharing agreement of some kind related to advertising.

To take just one example, The Guardian writes a lot of stories about soccer, but it can’t really target advertising to specific readers very well, since it is a mass-market newspaper. In other words, says Guardian developer Chris Thorpe, the newspaper fails to appeal to an Arsenal fan like himself because it can’t identify and target him effectively, and therefore runs standard low-cost banner ads. By providing the same content to a website designed for Arsenal fans, however, those stories can be surrounded by much more effectively targeted ads, and thus be monetized at a much higher rate — a rate the newspaper then gets to share in.

Open APIs and open platforms aren’t all that new. Google is probably the largest and most well-known user of the open API as a tool to extend the reach of its search business and other services, such as its mapping and photo services. Most social networks, such as Facebook and YouTube, also offer APIs for the same reason, though not all of them are as open as Google’s.

The Guardian, however, is the first newspaper to offer a fully open API (the New York Times has an API, but it doesn’t provide the full text of stories, and it can’t be used in commercial applications). It’s worth looking at why the paper chose to go this route, and what that might suggest for other companies contemplating a similar move — and not just content-related companies, but anyone with a product or service that can be delivered digitally.

For a content company like a newspaper, producing and distributing its content is the core of the business. Whether it’s in paper form or online, advertising usually pays the freight for the content, although subscription charges help, for both print papers and online versions like the Wall Street Journal, the Economist, etc. Many newspapers have regretted their decision to provide content online for free, since online advertising isn’t nearly as lucrative as print advertising (primarily because there are far more web pages to advertise on than there are newspaper pages, and therefore the supply outweighs the demand).

So why would a newspaper like The Guardian choose to provide access to its content via an open API, and not just some of its content, but everything? And why would it allow companies and developers to use that content in commercial applications? For one simple reason: There is more potential value to be generated by providing that content to someone else than the newspaper itself can produce by controlling the content within its own web site or service. You may be the smartest company on the planet, but you are almost never going to be able to maximize all the potential applications of your content or service, no matter how much money you throw at it.

As Thorpe described in a recent interview, the newspaper sees the benefits of an open platform as far outweighing the disadvantages of giving away content. By allowing developers to use the company’s content in virtually any way they see fit — and not just some of it, but the entire text of articles and databases the newspaper has put together — it can build partnerships with companies and monetize that content far more easily than it could ever do on its own.

This is effectively the opposite approach to the one that newspapers such as the Journal take, which is to up paywalls and charge users for every page they view, or charge them after a certain number of views (as the Financial Times does and as the New York Times is planning to do). It’s also the opposite approach to the one that companies like Apple take to their business — although Apple doesn’t produce content, it exclusively licenses and tightly controls the content it does handle (such as the music in iTunes), and it applies the same type of controls to its software and hardware.

Partnerships of the kind The Guardian is working on make a lot more sense for most companies that have lost the ability to control what happens to their content, something the Internet has done to virtually anyone whose product can be digitized and turned into bits, but has been particularly acute for content companies. By allowing others to make use of that content for their own purposes, and sharing in the revenue that comes from it, The Guardian takes what would otherwise be a disadvantage — the fact that it has lost control — and turned it to an advantage by becoming a platform. It’s a lesson other companies could stand to learn as well, instead of continually trying to reassert or recreate the control they have lost.

Twitter Annotations and the Future of the Semantic Web

Among the announcements at Twitter’s first “Chirp” conference for developers this past April was the launch of a new feature called Annotations. Unlike, say, “promoted tweets” or Twitter Places, Annotations aren’t so much a product launch as a substantial rethinking of the way the service functions on a fundamental level. The changes and extra dimensions it adds to Twitter could have a tremendous impact, not just on the social network and the developers and companies who make use of it, but on the way we interact with the web itself.

The new feature will be one of the first large-scale implementations of something called the “semantic web,” a term coined by World Wide Web Consortium director Sir Tim Berners-Lee. It refers to web technology with a built-in understanding of the relationships between its different elements — that is, everything from web pages to specific pieces of websites and services. Equipped with these kinds of tools, developers and companies can create applications and services that allow different pieces of the web to function together and exchange information, and therefore make services — from stocks to shopping to news — easier to use and more efficient.

An example of the semantic web used by Berners-Lee is the simple act of getting a cup of coffee with a friend. Instead of having to manage multiple different services or applications — calling or emailing the friend, checking a calendar, looking for a coffee shop nearby, checking a bus schedule — building semantic knowledge in would allow all of these different applications to talk to each other. You could simply choose a task in a specific piece of software, such as a calendar, and see dates and times that would work, as well as locations and bus routes automatically laid out for you.

While Annotations won’t make this high of a level of integration possible (at least not right away), the underlying principle is the same: Additional information, attached to an action, adds meaning to the behavior of users and can be interpreted in some way by software. The feature is expected to launch sometime later this year, and will allow developers to add that additional information to a tweet. That might include a keyword, a string of text, a hyperlink, a geographic location or virtually anything else that could be related to a message on the social network. These pieces of “metadata” won’t affect the character count of the original tweet, but will be carried along with it through the network and eventually be decoded, aggregated and filtered by a variety of applications or services (or by Twitter itself).

Twitter’s new feature isn’t the only large-scale experiment implementing the semantic web: Facebook is also rolling out its own version of metadata with its “open graph platform.” This involves an API as well as social plugins developers can add to web pages and services to allow users to “like” the pages they visit by clicking a button. Developers can then use the company’s open graph protocol to add metadata to this behavior, then track and filter that data in a variety of ways. For example, the site could detect that a user’s “like” occurred on a web page devoted to a movie, song or restaurant and track the most popular movies, etc. more easily.

Although Twitter and Facebook have both provided some guidelines for what kinds of activity and metadata they see developers and web sites integrating into their services, both social networks have also said that they will allow companies a substantial amount of leeway in coming up with their own ideas about what data to track or include.

The potential implications of this kind of semantic intelligence in social networks are substantial, because they will change the way we interact and use the web. A few examples include:

Reviews: Sites that involve restaurant, music or movie reviews could include metadata related to what a user is browsing when they post a comment to Twitter from a page, allowing other services to aggregate and filter that information to track popularity or make recommendations.

Stocks: Attaching a simple stock quote symbol to any tweet about a stock or a publicly traded company would allow services and users to track and aggregate information about those stocks, in the same way StockTwits does now.

Coupons: Companies could easily attach special offers to tweets that would be restricted to specific locations or specific times, allowing them to target users directly based on time or location.

Shopping: Metadata would allow sites to provide transaction info (if a user opted in) that would be attached to a tweet posted from a shopping site. This would also make it easy for services to rank and filter purchases, the same way Amazon does with its “people who bought X also bought Y” feature.

Music: Both users and services could track music-related tweets based on metadata involving the artist, genre, track, etc. Companies that want to target users based on a specific preference could then filter and analyze that data.

Games: Using metadata related to a specific social game, developers and companies could allow users to trade messages and play a form of reality game within a social network.

News: Any message that involved a current news story or location could have that information encoded in metadata, allowing users and services to track a developing story or event, as well as the conversation about it.

The impact of both Annotations and Facebook’s open graph protocol could turn out to be larger than either of those services individually: If services and applications that make use of one or both of these new technologies become popular with consumers, and the tools themselves become popular with developers, the semantic web envisioned by Sir Tim Berners-Lee and others could come closer to reality. That could change the way we interact with the web by making the software and services we use smarter and removing some of the friction between us and our social networks — and that will create new business opportunities not just for Twitter and Facebook, but for other smart technology companies as well.