Archives, algorithms and people

Post on 26-Jun-2015

244 views 0 download

Tags:

description

How we put the BBC World Service radio archive online using machines and crowdsourcing. A talk given to the UK Museums on the Web conference, November 2013. One of the major challenges of a big digitisation project is you simply swap out an under-used physical archive for its digital equivalent. Without easy ways to navigate the data there's no way for your users to get to the bits they want. We recently worked with the BBC World Service to generate metadata for their radio archive, 50,000 programmes from over 45 years. First using algorithms to generate "good enough" topics to put the archive online and then using crowd-sourcing to improve the data. Throughout 2013 we have been running this experiment to crowdsource improvements to the metadata that we automatically created. At http://worldservice.prototyping.bbc.co.uk people can search and browse for programmes, listen to them, correct and add new topics. This talk describes how we went about this and what we've learnt with this massive online multimedia archive - about understanding audio, automatically generating topics and crowdsourcing improvements to the data.

transcript

Tristan Ferne / @tristanfExecutive Producer

BBC Research & Development

Archives, algorithms and peopleor

How we put the BBC World Service radio archive online using machines and

crowdsourcing

The BBC World Service archive

1947-2012

Spelling mistake

Missing data

Sometimes incorrect dataNo semantic data

The missing metadata

How it works

Listening machines

Noisy transcripts

Algorithms

Algorithms and people

The prototype

Show Synopsis editing version

Machine learning

Results

70000tag edits

How much data?

1000synopsis edits

71000edits

36000listenableprogrammes

1mmachine tags

70000programmes

3000users

of programmes listened to36%

of programmes tagged21%

And four lost programmes

Tags are a large and sparse space

When is a tag correct?

When is a programme tagged completely?

How do you measure crowd-sourced data?

How good is the data?

Who does the work?

1 person = 30% of edits

10 people = 70% of edits

10% of people = 98% of edits

The shape of the archive

Places mentioned

Linking from the News

The Last Danish Christmas Broadcast

“Entirely in Danish”

We can significantly improve the data

It’s cost-effective with re-usable technology

A crowdsourcing approach

What we’ve learnt

How good are the machine tags?

How much crowdsourcing do you need?

When is your data good enough?

Open questions

worldservice.prototyping.bbc.co.ukwww.bbc.co.uk/rdgithub.com/bbrd

tristan.ferne@bbc.co.uk@tristanf