Categories
Collection databases Conferences and event reports Folksonomies Web 2.0 Web metrics

Web Directions South 2007 – presentation and some thoughts

Web Directions South 07 was lots of fun and there were some great presentations over the two days. Unfortunately conferences are always full of choices and I missed several presentations I’d been looking forward to catching. That said, overall the quality was high and there were only a handful of dull moments. Most of the presentations I saw were not on the tech-side (JS, Ajax, CSS etc) of things – Luke was there to go to those.

Here’s some notes from my highlights.

Cameron Adams managed to pack out one of the smaller rooms and by the time his ‘Future of web based interfaces’ was in full flow there were about 50 people standing at the back. Adams’ presentation went through the possibilities of flexible interfaces that are both customisable by the user (much like Netvibes or iGoogle is) and automatically reformats as you use it (like the BBC News pages subtly do).

After my own presentation (see below) it was on to Scott Gledhill’s ‘Is SEO evil?‘ to which the answer is, of course, no. SEO and a web standards approach should be complimentary. Scott had some lovely images – the menacing gummi bears in particular – and a fascinating case study from News Digital Media around the Steve Irwin death. In this instance, News went out with a web headline that was far more immediate and keyword loaded (“steve irwin dead”) than their major competitor, Sydney Morning Herald/Fairfax who were more obscure (“crocodile man reported dead”). They tracked the story traffic and referrers by the hour and more than doubled the Fairfax traffic – even after Fairfax adjusted their headline. Scott also told how journalists are now much more SEO content-savvy in their writing and that his team gives the journalists the necessary web reporting tools to be able to track their own stories. This, combined with the highly competitive environment, encourages journalists to further refine and re-edit their stories for performance even after initial publication.

The second day began with an edit of Scott Berkun’s famous Myths of Innovation presentation. Scott’s main message is that you can’t force ‘innovation’ and that it needs time and space to happen organically. In fact, one of the best triggers of innovation are failures and mistakes. He suggests that perhaps we should start including a ‘failures’ budget line in our organisational budgets – accept that they will happen and that we are all the better for it.

George Oates from Flickr spoke about how Flickr manages and facilitates user communities. She started out tracing Flickr back to its origins at Ludicorp as a sort-of MMORPG called Game Neverending. After GNE folded the community that had grown around it was imported directly into Flickr and they brought their experiences from the game world into the construction and design of Flickr. I found her focus on users and the real need for human-to-human communication and relationship management that Flickr does a timely reminder that in the museum world we cannot expect communities to ‘just happen’ around our content and that when the seeds of community appear they need careful nurturing. The necessary nurturing is impossible if you move immediately on to the next project.

Adrian Holovaty, the mind behind Chicago Crime and several other datamining and visualisation projects gave a fascinating presentation about the hidden potential of structured data. Now over in the museum world we are experts at structured data but we rarely make the most of it. Throughout Holovaty’s talk he kept coming back to the ideas of serendipity and free browsing that I’ve been working on with our OPAC. His position was to make everything hyperlinked and let the users build their own paths through the data. To that end he built the Django Databrowse application which takes a database and basically build a simple website that allows users to link from anything to anything else. Following Chicago Crime which took flat datasets from the Chicago Police Department and made them navigable in ways that the Chicago PD had never intended (view crimes by area, visualise hotspots, map your jogging route against reported crimes etc), Holovaty went on to do some great visualisation work at the Washington Post. Here he asked journalists to enter their notes into a simple database as well as turning their notes into stories. This allowed him to build the Faces of the Fallen which tracks and maps every US soldier killed in Iraq. Faces not only reveals some uncomfortable patterns in the data (deaths by age of soldier, by state etc), it also has allowed linkages to family tributes and newspaper articles about the circumstances of their death. The project returns great value back to reporters and the paper who can now report ‘milestones’ and trends, but also to the community who can now make ‘more sense’ out of what would otherwise be simply seen as a list of names. It ‘humanises’ the data, giving it far greater impact. Holovaty is now working on a community news project Every Block which intends to harvest and aggregate content by ‘block’ from various news sources – automatically creating journalistic stories from raw data. (Reuters already does this with some financial reporting).

There are a growing selection of presentation slides over at Slideshare.

Here’s an edited version of my own presentation slides which use the Powerhouse Museum’s collection search and tagging implementation as a case study of a government implementation of Web 2.0 techniques. Those who have seen my presentations over the recent months will recognise some re-use and re-puposing. For various reasons I have had to remove about 20-30 slides but most of it is there. There is a podcast coming apparently.

Categories
Developer tools Imaging Web 2.0

Comparing a site across browsers

One of the biggest problems when designing and developing a new website or rolling out a new look and feel is cross-browser compatibility. Usually the solution has been to have a series of machines, real or virtual, with different versions of the various different browsers out there installed, and then go through each one laboriously.

Fortunately now there is Browsershots which is a web-based browser farm which you can utilise for browser checking. You simply submit an URL to Browsershots and tick the various flavours and versions of browser you want to check against and then wait . . .

Your request is queued and once processed you can view and download screenshots of your site as it look in each of the browsers selected. Because it runs on actual machines these screenshots aren’t kludges, they are the real thing.

Categories
Conceptual Digitisation

Subscription museum content? Some implications of the NYT announcement for museums

The New York Times announced last week that it was stopping charging for archival and subscription content on its website. As its self-report explains, the NYT has realised that selling and managing subscription service to archival content now is not going to be as profitable as selling advertising on this content and making access to it easy and free.

This turnaround has come at the hands of Google and the power of search – search which is now driving less ‘serious’ readers to their content who are unwilling to buy a subscription. For the NYT they expect to get better returns on material previously only available through subscriptions via onsite advertising and the ability for readers to engage in conversations around the content. Conversations mean exposure, exposure means advertising. Blogger Jason Kottke (amongst many others) has already been digging through the archives exposing some of the more interesting material.

There has been an explosion of discussion across the web but the Future of the Book pulls together three commentaries to suggest that

Whatever the media business models of tomorrow may be, they will almost certainly not revolve around owning content.

and drawing on Jeff Jarvis’s 2005 proclamations, explains that the future lies in integrating your content into your audience’s conversations.

But in this new age, you don’t want to own the content or the pipe that delivers it. You want to participate in what people want to do on their own. You don’t want to extract value. You want to add value. You don’t want to build walls or fences or gardens to keep people from doing what they want to do without you. You want to enable them to do it. You want to join in.

So what might this mean for museums?

Whilst those in collecting institutions often see themselves as the sole holders of particular objects, it is becoming more acceptable to acknowledge that as institutions there is still a lot to learn about these objects and that that knowledge may lie elsewhere in the community. Certainly injecting museum content into, and encouraging audience conversations then is not as controversial as it might have been ten years ago. Exclusivity might work with our physical spaces but not online.

At the same time increasing commercial pressures are asking museums to find new revenue streams – image sales, licensing, syndication, partnerships. Already the V&A and the Met have moved to ‘no-fee’ image licensing for small run academic publishing after discovering that the internal cost of charging for these operations outweighed their commercial returns. From the WIPO’s Guide to Managing Intellectual Property for Museums (pt 6.6) –

Recent developments in business models concerning the production and distribution of content on the Internet, coupled with a continued examination by museums of their missions and mandates has led to an awareness that the making available of museum images is merely a means to a commercial end, and not the end, itself. Indeed, in a recent press release, the Victoria and Albert Museum announced that it would no longer charge fees for academic and scholarly reproduction and distribution for its images, claiming that while it earned approximately 250,000 a year from scholarly licensing programs, the overhead costs associated with licensing fees rendered their profits much less. What is not reported, but suspected, is that the Victoria and Albert Museum determined that it was wise business practice to allow its copyright-protected images to be made available for free, thereby increasing their circulation and delivering significant promotional opportunities back to the museum.

As the WIPO Guide suggests, there is some potential for museums in the online space in the brand and promotional opportunities in the short term, and then flowing from these in the medium-long term, the partnerships and commercial content syndication options that are expected to flow from a greater awareness (and discussion) of content.

What the NYT announcement does, along with the increased commercial activity around digitising state-held records, especially those relating to the profitable family history space, is create significant competition for museum content online.

Categories
Social networking Web 2.0

Who participates? The many and the few

Who is really participating on so-called ‘social media’ websites? Research at both the academic and the market research level continues to probe this question.

Firstly, there is a fascinating paper titled ‘Power of the Few vs. Wisdom of the Crowd: Wikipedia and the Rise of the Bourgeoisie ‘ by Kittur, Chi, Pendlteon, Suh and Mytkowicz presented at SigGraph 2007 that looks at the changes in contribution trends to Wikipedia and charts the declining influence of an elite cabal of admins, and then compares this to behaviour on del.icio.us.

In this paper, we show that the story is more complex than explanations offered before. In the beginning, elite users contributed the majority of the work in Wikipedia. However, beginning in 2004 there was a dramatic shift in the distribution of work to the common users, with a corresponding decline in the influence of the elite. These results did not depend on whether work was measured by edits or by actual change in content, though the content analysis showed that elite users add more words per edit than novice users (who on average remove more words than they added).

This paper’s research contributes to Wilkinson and Huberman’s ‘Assessing the value of cooperation in Wikipedia‘ at First Monday (from April) which asserts that there is a significant correlation between article quality and number of edits.

Then over at the McKinsey Quarterly this month (requires a free registration), McKinsey presents their comparatively small-scale research (573 users) from Germany to examine some of the potential lessons from why Germans upload content to video sharing sites.

Few companies, however, have a clear understanding of what inspires users to contribute to such sites. Executives might start by looking to the world of online video sharing, another fast-growing test bed for participation. McKinsey research conducted in Germany finds that motives such as a desire for fame and a feeling of identification with a community encourage collaboration and participation. Such findings, we believe, offer insights into the way companies might tailor their Web 2.0 offerings.

Categories
Web 2.0 Young people & museums

Brief notes on the vernacular web, class and design

A few days I was on Facebook and saw an advertisement to add ‘glitter’ to my profile. And then I came across Russian net artist/curator Olia Lialina’s highly entertaining essay called Vernacular Web 2 (which comes complete with an ‘almost spam’ URL).

Lialina’s exploration of the new forms of ‘vernacular’ design explores the transition from ‘My Home Page’ to customised ‘services’ (iGoogle, MySpace, Facebook etc).

Categories
Collection databases Interactive Media Young people & museums

C is for collection – an ABC book with collection objects

Two weeks ago we made a simple ABC book for young children available on our children’s website. It is called ‘C is for collection‘ and is a very basic extension of our online collection built in Flash with an XML file supplying the necessary collection data allowing for easy expansion.

A longer term objective of ‘C is for collection’ is to build a database of child-friendly object descriptions and explore the options for a children’s tagging game with the same XML.

Have a play and remember to turn up your speakers. More children’s games are coming shortly.

Categories
Folksonomies Web 2.0

Steve.museum update as a podcast

I’ve just finished a presentation to art museum folk at the Sites of Communication 3 conference at the National Gallery of Victoria, and true to form there was quite a bit of interest in social tagging. There seems to now be widespread awareness of the problem of the ‘semantic gap’ between the language of art museums audiences (especially as they are being seen to be diversifying) and that of art curators and researchers. And there is increasing interest addressing this problem.

Thus when museum people ask about collection tagging projects other than our own, I send them off to the Steve.museum project website. Invariably they come back, having dipped their toes into some of the research material, with more questions. Jennifer Trant has produced a rather excellent podcast summary of the project to date and some of the preliminary results emerging from it. The podcast is a good example of making what is otherwise a time consuming and text heavy task an easy-to-digest and informative 12 minute presentation – complete with a few slides. (It uses the M4A format so you will need Quicktime or iTunes.)

Steve.museum is doing some excellent and very considered research that will re-assure many tag skeptics and no doubt lead to more and better tagging implementations down the track. Whether the Steve results will be able to be applied directly to collections outside of visual art – social history and natural history collection especially – remains to be seen.

Categories
Conceptual Social networking Web 2.0 Web metrics Young people & museums

Social production, cut and paste – what are kids doing with ‘your’ images?

It has been one of the worst kept secrets of web statistics – deep linked image traffic. While this has been going on for years, since the beginning of the WWW actually, it has increased enormously in the past few years. On some cultural sector sites such traffic can be very substantial – a quick test is to look at exactly how much of your traffic is ‘referred’ from MySpace. It is also one of the main reasons why Photobucket has traditionally reported traffic so much higher than Flickr is – its deep linking and cut and paste engagement with MySpace. With the move away from log file analysis to page tagging in web analytics, some, but not all of this deep linking traffic is fortunately being expunged from analytics reporting.

Two Powerhouse examples include a Chinese news/comment portal that deep linked a Mao suit image (from an educational resource on our site), sending us 51,000 visits in under 24 hours in August 2005, and an A-grade Singaporean blogger who deep linked an image of Golum (from our archived Lord of the Rings exhibition pages) to use to describe an ugly celebrity which generated over 180,000 visits over 8 days In January 2007. (In both of these examples the visits were removed from the figures reported to management and funders.)

What is going on here sociologically?

At the recent ICA2007 event in San Francisco danah boyd and Dan Perkel presented an interesting look at the subcultural behaviours that are, in part, producing this effect. Although they look specifically at MySpace there are threads that can be drawn across many social sites from forums to blogs. Drawing on the work of many cultural theorists, they argue that on MySpace what is going on is a form of ‘code remix’. That is, young people’s MySpace pages are essentially ‘remixes’ of other content – but unlike a more traditional remix in audio and video cultures, these code remixes occur through the simple cut and paste of HTML snippets. By ‘producing’ both their MySpace pages as well as their online cultural identity in this way, they are reshaping concepts of ‘writing’ and digital literacy. They are also, importantly, not in control of the content they are remixing – a deep linked image can easily be changed, replaced or removed by the originating site.

There are plenty of examples – boyd and Perkel give a few – where the content owner changes the linked image to disrupt the deep linker. In the case of our Singaporean blogger we renamed the linked image to prevent it from appearing on her site (and in our statistics).

Revealingly, Perkel’s research is showing that many MySpace users have little, if any, knowledge or interest in website production – that is CSS and HTML. Instead, what has formed is a technically simple but sociologically complex ‘cut and paste’ culture. This is what drives the ‘easy embedding’ features found on almost any content provider site like YouTube etc – it is in the content providers’ interest to allow as much re-use of their content (or the content they host) because it allows for the insertion of advertising and branding including persistent watermarking. Of course, the museum sector is not geared up for this – instead our content is being cut and pasted often without anyone outside the web team having a deep understanding of what is actually going on. There are usually two reactions – one is negative (“those kids are ‘stealing’ our content”) and the other overly positive (“those kids are using our content therefore they must be engaging with it”). Certainly Perkel and others research deeply probelmatises any notion that these activities are in large part about technical upskilling – they aren’t – instead those involved are learning and mastering new communication skills, and emerging ways of networked life.

One approach that some in the sector have advocated is the widget approach – create museum content widgets for embedding – to make repurposing of content (and code snippets) easier. There have been recent calls for museum Facebook apps for example. But I’m not sure that this is going to be successful because a great deal of embeds are of the LOLcats variety – perhaps trivial, superficial, but highly viral and jammed full of flexible and changing semiotic meaning. Whereas our content tends to be the opposite – deep, complex and relatively fixed.

Categories
Digitisation Web 2.0

How to do low cost transcription of hand written and difficult documents

So your museum has already done the easy part of digitisation – taking digital photos of your objects, but now you have a complex hand-written materials you need to digitise . . . what can you do?

This is a question that has popped up in several meetings over recent months.

Our Curator of Information Technology, Matthew Connell, came up with a brilliantly simple solution – and there is no need for the original material to leave your organisation.

With the low cost of MP3 recorders it is very to now record large amount of audio into a single file, already compressed. Take one of these MP3 recorders and ask the expert who is familiar with the document or material requiring digitisation to read the document clearly into the recorder. This may be done over an extended period of time – there is no need to do it all in one go.

When completed, upload the MP3 of clearly spoken audio to a web server. Then use one of several online audio transcription services to transcribe the audio. We have been using such services to get quick, low cost transcriptions of public lectures and and podcasts, and have been impressed with their timeliness and accuracy.

Even factoring in the cost of reading time, this will almost certainly be cheaper and more error free than scanning and transcribing directly from the written original. It also provides significantly more flexibility in terms of pricing as there is a high level of competitiveness amongst audio transcription services at the moment – a level of competition that may not exist amongst specialist written services.

Categories
Conceptual Digitisation

Filtering memory – SEO, newspaper archives, museum collections

When Bad News Follows You in the New York Times (via Nick Carr) is a fascinating article about what can happen when ‘everything’ is put online.

The article looks at the new array of problems that have come about as a by-product of the NYT optimising their site and archives for Google with SEO techniques. Suddenly stories that were either of minor significance, or were in later editions, corrected, are appearing toward the top of Google searches for names, places and events.

Most people who complain want the articles removed from the archive.

Until recently, The Times’s response has always been the same: There’s nothing we can do. Removing anything from the historical record would be, in the words of Craig Whitney, the assistant managing editor in charge of maintaining Times standards, “like airbrushing Trotsky out of the Kremlin picture.”

Whitney and other editors say they recognize that because the Internet has opened to the world material once available only from microfilm or musty clippings in the newspaper’s library, they have a new obligation to minimize harm.

But what can they do? The choices all seem fraught with pitfalls. You can’t accept someone’s word that an old article was wrong. What if that person who was charged with abusing a child really was guilty? Re-report every story challenged by someone? Impossible, said Jonathan Landman, the deputy managing editor in charge of the newsroom’s online operation: there’d be time for nothing else.

(snip)

Viktor Mayer-Schönberger, an associate professor of public policy at Harvard’s John F. Kennedy School of Government, has a different answer to the problem: He thinks newspapers, including The Times, should program their archives to “forget” some information, just as humans do. Through the ages, humans have generally remembered the important stuff and forgotten the trivial, he said. The computer age has turned that upside down. Now, everything lasts forever, whether it is insignificant or important, ancient or recent, complete or overtaken by events.

Following Mayer-Schönberger’s logic, The Times could program some items, like news briefs, which generate a surprising number of the complaints, to expire, at least for wide public access, in a relatively short time. Articles of larger significance could be assigned longer lives, or last forever.

Mayer-Schönberger said his proposal is no different from what The Times used to do when it culled its clipping files of old items that no longer seemed useful. But what if something was thrown away that later turned out to be important? Meyer Berger, a legendary Times reporter, complained in the 1940s that files of Victorian-era murder cases had been tossed.

“That’s a risk you run,” Mayer-Schönberger said. “But we’ve dealt with that risk for eons.”

There are interesting parallels with our experience in making our online collection more usable and accessible. Public enquiries have skyrocketed and now range from the scholarly to the trivial – the greatest increase being in the latter category. Whilst there is a significant amount of extremely valuable piece of object related information sent in by members of the public, there are false leads and material that cannot be adequately verified, and more still that the Museum already knows but has not yet made available online. Managing public expectations and internal workflow is a difficult balancing act and a continuing challenge that many museums that not only put their collections online, but also make them highly accessible, are facing.