JISC09 Open Access

This session on the economic impact of Open Access recently published. John Houghton who authored the report with Charles Oppenheim is going to talk about the report to start with.

The project tried to quantify the costs and benefits – creating a series of spreadsheets contains elements identified in the process model of Scholarly Publishing, adding in cost data. There are about 2300 activity items that are costed in these sheets. Some example figures for activities in UK Scholarly publishing in 2007 – Reading cost £2.77billion, writing £1.6billion, Peer review £203million etc.

The overall estimate (for UK scholarly publishing in 2007) was £5.4billion

Then looked at cases and scenarios exploring cost savings result from the alternative publishing models throughout the system. Finally models the impact of changes in accessibility and efficiency on returns to R&D.

In summary – OA publishing models (whether Author pays, Overlay etc.) should save money.

John says ‘of course there would have to be a move of money from subscriptions to e.g. author pays funds’ – lets not underestimate the impact of this – this move of funds is likely to be politically charged, and challenging to organisations. I would also be interested in seeing some analysis of how (for example) ‘author pays’ might change the profile of expenditure across institutions – would this result in expenditure being more or less concentrated in research heavy institutions, or is it neutral in this respect?

Other side of the coin – benefits of Open Access models also more than benefits of traditional publishing.

See http://www.cfses.com/EI-ASPM/ for the opportunity to see, and play with, a simplified model.

Now Hector MacQueen (from University of Edinburgh School of Law) talking about Legal Perspectives on OA Publishing – going to make remarks from personal experience. Hector started by thinking about doing research by electronic means – and only gradually come to think of it in terms of Open Access.

In 1978 research was based around physical access – Hector’s material in libraries and archives. You had to get yourself to the physical location, or sometimes via ILL (although often material Hector wanted was not available via ILL).

The first electronic resource in Hector’s area was Lexis (now Lexis-Nexis) – but it was made very clear to the academics that this was not free. Not only was Lexis restricted to a single terminal but there was a ‘gatekeeper’ (person) who you had to go to to get searches done.

Courts (and others) started to make material available on the web – for free. So for formal sources, this was the start of ‘open access’ – or at least free access. The library became less of the place to get resources – moved to the desktop.

Then Hector started exploring the website of individuals who were publishing – and became aware of ‘self-archiving’ activity – especially in the USA. Hector followed suit. There was already a lot of informal sharing (and what Hector describes as ‘informal peer-review’ – essentially pre-publication comments from peers) happening – via email etc.

Hector notes that the Open Access world needs to recognise this informal activity. I agree – I’d go further and say that one of the problems we (Libraries/OA Movement) have is that we tried to formalise this type of activity, rather than working to support the existing informal sharing. Hindsight is a wonderful thing, but I’m not sure that we are past this stage yet (certainly not in all disciplines) – a consideration of how we can support informal sharing woudl still be a valuable exercise I think.

Hector now commenting on Copyright – the impact of the Google Book Search agreement (how will this impact in the UK?), Project Gutenberg, European Digital Library, Amazon + Kindle – also noting the impact of iTunes on availability of music and the dropping of DRM.

Finally before group discussion, an academic (who?) in Theatre studies.

Those studying early theatre groups (e.g. The King’s Men – Shakespeare’s troop) – have problems tracking records, as can be extremely distributed (around county record offices, various sources in their home location etc.) There is work (Reed) bringing together records from all over the UK – which are being published in an expensive series of books – one per town, a series that has been updated over the last 30 years (I think that’s right). However, the leader of the project negotiated from the off (30 years ago) that he had personal rights to the digital distribution – and so he can now make all the pdfs for the publications available via the Internet Archives, and build a database of information listing actors, troops, locations, plays, writers.

EEBO is another source – a resource created from this called DEEP (Database of Early English Playbooks) – which uses book title pages from EEBO to build a database of Early English Play Books.

EEBO is available via a JISC Collections deal in the UK. This, combined with seamless authentication via IP address, leads people to believe it is ‘free’ – this is an issue when trying to get people to understand the importance of Open Access.

The academic talking about a journal he is involved in which is not Open Access and reflecting why – an editorial stipend from the publisher allows the academics involved to promote the work. However he suspects that if these costs were met effectively by the tax payer, then the overall cost would be lower, as the

Q: I think the question was: What about Open Access outside Higher Education – that is should OA material created by HE be available to all others or just HE community?

A: In general the debate around OA has been focused on STEM – but OA would have a big impact on e.g. Law firms as well

A: The pharmaceutical industry could be a key beneficiary of OA – this is not necessarily a problem but needs to be recognised

A: Need to consider carefully how institutions add value – perhaps what is published is not the value add, but the interpretation of that – knowledge transfer etc.

Q: What would we need to do to make researchers fill the repositories that are out there and empty?

A: (not sure who said this, but a researcher) Mandating deposit essential – but not popular with the academics. Realising that some academics don’t want to be read!

A: Mandating good thing (John Houghton). Something that persuades colleagues is finding things that are available via OA. This means that your own institutional repository is very rarely of interest – it is what is being published elsewhere.

A: Charles Oppenheim – Citation advantage compelling – OA material gets cited more, but the reasons for this not well understood. University of Southampton now more highly cited than Oxford/Cambridge (according to Southampton)

Interesting that the view from the researchers is ‘mandate’. Really good to get researcher’s views, but suspect that the reason they are speaking at this event is exactly because they are atypical.

Technorati Tags:

JISC09 Keynote

The keynote is from Professor Lizabeth Goodman who is Prof of Creative Technology Innovation at the University of East London and Director of SMARTlab Digital Media Institute and MAGIC Multimedia and Games Innovation Centre – and also now involved in FutureLab (not sure what capacity)

In the current economic environment Lizabeth has found she is being asked to talk more at conferences about ‘forget the money, think about the people’ type stuff.

The keynote dotted through a lot of the work and projects from SMARTLab, and probably the best place to direct you is http://www.smartlab.uk.com/2projects/index.htm – it all sounds really interesting, but I didn’t find the presentation that engaging I’m afraid – too much reliance on a video and talking over it – anyway here are some brief notes.

Lizabeth playing a video about the work at SMARTLab –  one motto is “JUST MAKE IT” – the money always follows!

Now showing a clip from an early project – the Interactive Shakespeare – on CD-ROM  allowing the viewers to recut the film – participants people like Fiona Shaw, RSC etc. Had to work out how to engage 6000 Open University students and many more casual viewers.

I’m not sure I’m getting this – Lizabeth talking about our relationship to the material and ‘kinaesthetic’ experience – but I feel like I’m coming into a lecture half way through – I don’t understand half of what she is say I’m afraid.

Lizabeth is saying is that people who move in different ways (e.g. those with severe physical disability) learn in different ways – I think. Some work on using motion tracking to provide animated representations of the participants – allowing them to create different versions of themselves – an Avatar before there were virtual worlds such as Second Life readily available.

The Trust Game – allowing children to interact with the game by changing the form of the characters etc.

A project at the Stephen Hawking school in East London. Wanted to learn about the learning of severely disabled children. Which has informed the next project ‘Interfaces’. This uses a (large) screen that tracks eye movement (MyToby) – originally developed in the commercial sector for selling, was adapted to create an interface where you use your eyes to control the interface to create music.

Now relating how abused women were able to tell stories through pictures when they would never been able to tell the stories through speech – Safetynet project.

Chick2Go – where women with disabilities are mapping the streets of East London for the 2012 Olympics and Paraolympics – looking at safety and accessibility issues – where are there wheelchair ramps etc.

Some work in SecondLife – Wheelies – a wheelchair danceclub – everyone in the virtual space has to use a wheelchair

Stressing the importance of interacting with things and people that are close by, rather than remote (dissing Twitter – this is silly, it is just a technology, it is neutral on location – and since it contains location information, it can be used to identify and interact with local resources and people)

Project with Microsoft called ‘ClubTech’ – providing technology to US childrens clubs.

Bits and pieces – Fizzy, MobiMissions

Technorati Tags:

Opening Digital Doors – JISC09

Today I’m in Edinburgh for the JISC Conference. I’ve already done a brief post on a demonstration of the SWORD API for depositing material into digital repositories, but now the opening session with an introduction by Malcolm Reid (Executive Secretary JISC), and address by Tim O’Shea (Chair of JISC).

Malcolm Reid starts by reflecting on some of the issues raised yesterday at the ‘pre-conference’ sessions held yesterday, which unfortunately I wasn’t able to attend. This is a bit of a mixed bag – I think I might recommend looking out blog posts from yesterday instead of trying to summarise here.

Now Tim O’Shea talking about what JISC has been, is, and will be doing:

  • Research Excellence
    • Virtual Research Environments
    • Re-use of Research Materials in E-learning environments
  • Institutional Drivers
    • Open Access Learning Materials
    • Knowledge Transfers – engagement with commercial environment
  • Making Savings and Working Smarter
    • Access Management – UK Federation now the largest Access Management Federation in the world!
    • JISC Collections
      • Saviing money on license agreements
      • Giving access to more resources
    • Institutional Repositories
    • Green ICT – huge numbers of computers in UK HE – cost to sector over £100million in ICT electricity bills (I’d be suprised if this was anything like an accurate figure to be honest – instinctively feels low, don’t ask me to justify that)
  • Next 12 months
    • New strategy for JISC for 2010-12
    • Focus on Value for Money
    • Looking at Enterprise-wide Systems
    • Looking at Knowledge Transfer and Wealth Creation

Highlights for Tim over the day:

  • Launch of Box of Broadcasts (BoB) by BUFVC
  • New/updated services from Edina and Mimas (Geocode crosswalk, Archives Hub and COPAC)
Technorati Tags:

SWORD

Adrian Stevenson from UKOLN talking about, and demonstrating, SWORD. Uses Atom Publication Protocol (APP) profile (with some custom extentions) to provide a standard way of interacting with digital repositories such as DSpace, Fedora, E-Prints, Intralibrary etc.

The SWORD work has developed the APP profile, and paid for work to enable SWORD deposit to repository software in common use in HE. It has also enabled the development of some demonstration clients:

  • http://client.swordapp.org is a simple web client
  • Netvibes SWORD widget (available along with other JISC demonstrator widgets from http://rwidgets.co.uk) – example of importing image into Intralibrary – some issues that still need tidying up – e.g. resource name once loaded into Intralibrary is currently just a string of alphanumberic characters
  • Facebook App – (called SWORDAPP)
  • FeedForward – this is a tool being developed by staff at CETIS – allowing you to pull together lots of resources from various sources (focussing on RSS but also allowing you to add local files and other resources) – it will package up a set of resources in an appropriate packaged format for the target repository (e.g. METS for DSpace, IMS for Intralibrary)

There are some issues – but I’m not clear from a quick demonstration whether these are things that need further development in SWORD or just to do with the way the apps are currently coded. One of the issues is choosing appropriate metadata and deposit types, and a question of how the App might retrieve these from the target repository so it can prompt the user for the correct values.

 

Technorati Tags: ,

Its time to change library systems

Recently Chris Keene (University of Sussex) sent an email to the LIS-E-RESOURCES email list about the fact that in academic libraries we are now doing a lot more ‘import’ and ‘export’ of records in our library management systems – bringing in bibliographic records from a variety of sources like book vendors/suppliers, e-resource systems, institutional repositories. He was looking for some shared experience and how other sites coped.

One of the responses mentioned the new ‘next generation’ search systems that some libraries have invested in, and Chris said:

“Next gen catalogues are – I think – certainly part of the solution, but only when you just want to make the records available via your local web interface.”

One of the points he made was that the University of Sussex provides records from their library management system to others to allow Union catalogues to be built – e.g. InforM25, COPAC, Suncat.

I sympathise with Chris, but I can’t help but think this is the point at which we have to start doing things a bit differently – so I wrote a response to the list, but thought that I’d blog a version of it as well:

I agree that library systems could usefully support much better bulk processing tools (although there are some good external tools like MarcEdit of course – and, scripting/programming tools (e.g. the MARC Perl module) if you have people who can programme them. However, I'd suggest that we need to change the way with think about recording and distributing information about our resources, especially in the light of investment in separate 'search' products such as Aquabrowser, Primo, Encore, Endeca, &c. &c.

If we consider the whole workflow here, it seems to me that as soon as you have a separate search interface the role of the 'library system' needs to be questioned – what are you using it for, and why? I'm not sure funnelling resources into it so they can then be exported to another system is really very sensible (although I absolutely understand why you end up doing it).

I think that once you are pushing stuff into Aquabrowser (taking Sussex as an example) there is little point in also pushing them into the catalogue – what extra value does this add? For books (print or electronic) you may continue to order them via the library system – but you only need an order record in there, not anything more substantial – you can put the 'substantial' record into Aquabrowser. The library system web interface will still handle item level information and actions (reservations/holds etc.) – but again, you don't need a substantial bib record for these to work – the user has done the 'searching' in the search system.

For the ejournals you could push directly from SFX into Aquabrowser – why push via the library system? Similarly for repositories – it really is just creating work to covert these into MARC (probably from DC) to get them into your library system, to then export for Aquabrowser (which seems to speak OAI anyway).

One of your issues is that you still need to put stuff into your library system, as this feeds other places – for example at Imperial we send our records to CURL/COPAC as well as other places – but this is a poor argument going forward – how long before we see COPAC change the way it works to take advantage of different search technology (MIMAS have just licensed the Autonomy search product …). Anyway – we need to work with those consuming our records to work out more sensible solutions in the current environment.

I'd suggest what we really need to think about is a common 'publication' platform – a way of all of our systems outputting records in a way that can then be easily accessed by a variety of search products – whether our own local ones, remote union ones, or even ones run by individual users. I'd go further and argue that platform already exists – it is the web! If each of your systems published each record as a 'web page' (either containing structured data, or even serving an alternative version of the record depending on whether a human or machine is asking for the resource – as described in Cool URIs), then other systems could consume this to build search indexes – and you've always got Google of course… I note that Aquabrowser supports web crawling – could it cope with some extra structured data in the web pages (e.g. RDFa)?

I have to admit that I may be over estimating how simple this would be – but it definitely seems to me this is the way to go – we need to adapt our systems to work with the web, and we need to start now.

The Future is Analogue

"Far out in the uncharted backwaters of the unfashionable end of the western spiral arm of the Galaxy lies a small unregarded yellow sun.

Orbiting this at a distance of roughly ninety-two million miles is an utterly insignificant little blue green planet whose ape-descended life forms are so amazingly primitive that they still think digital watches are a pretty neat idea."

Douglas Adams, The Hitchhiker's Guide to the Galaxy

Digital Libraries, Digital Repositories, Born Digital, Digital Objects – the idea of digital information has become an intrinsic part of the library landscape in the 21st century. However, I believe that as we manage more information in digital formats, we need to think about managing it in analogue, rather than digital, ways.

What do I mean by 'digital' and 'analogue' in this context? Well – to be clear, I'm in favour of using computers to help manage our data – in fact, I think this is key to our ability to take an 'analogue' approach!

Digital values are absolute – something is either on or off, 1 or 0, black or white. Analogue values live along a continuous scale – from black to white and all the shades of grey in between. Computers store information as a series of bits – which can either be on or off – there is no grey here, a bit is either on (1) or off (0) – they are literally digital.

When dealing with physical items on a shelf, and entries in a printed or card catalogue, it is difficult to do anything but take a digital approach to managing your library – something is either on this shelf, or that shelf; on this card or that card; about this subject or about that subject.

Even now we don't rely on printed/card catalogues, and many items are available in electronic, rather than physical, format, we are still managing our collections in this 'digital' way. We treat all information in our catalogues as 'absolute' – from titles to subject headings.

I've heard Tim Spalding of LibraryThing talk about this in terms of subject headings – he said 'somebody wins' when you assign subject headings in a traditional library catalogue.

Even questions of fact, which you'd generally expect to have a single answer may not be entirely 'digital' (right or wrong). The classic example used in library school for reference questions is 'how high is Mount Everest?' – if you check several reference works you may come up with several answers – Wikipedia covers some of the various answers and why they are different.

At this point you may be wondering what the alternative is – you've still got to allocate a subject heading at some point (assign a title, author etc.) – right? Well, I think the answer in one of the most effective mechanisms for storing and retrieving information we've got – the web.

What makes the web 'analogue' rather than 'digital' in the way I'm using the terms is the link. We can see this clearly in the way Google was originally designed to work. In "The Anatomy of a Large-Scale Hypertextual Web Search Engine" Sergey Brin and Larry Page describe how Google was designed to make use "of both link structure and anchor text".

As is well known, Google uses the concept of the 'Page Rank', which is calculated based on the links between pages, but as illustrated by this diagram, it isn't a straightforward count of the number of links to a specific page, but allows for different weights to be assigned to the links

725px-PageRanks-Example 

You can see that E has many more links than C, but does not get such a high page rank as it is, in turn, not linked to by any high ranking pages.

The Page Rank gives some kind of 'authority' to a page, but then there is the question of what the page is actually about. This latter question is not simple, but one factor that Brin and Page were explicit about is that "The text of links is treated in a special way in our search engine … we associate it with the page the link points to"

This means that not only is each link a 'vote' for a page in terms of page rank, but that it is also a piece of metadata about the page it is linked to. If you look at all the text of each link used, you are bound to get a wide range of text – as different people will link to a page from different perspectives – using different terminology and even different languages.

Suddenly here we are thinking about a way of classifying a document (web page) that allows many, many people to participate – in fact, as many people as want to – the architecture of the web puts not limit on the number of links that can be supported of course.

Alongside each assertion of a description also has a weight associated with it – so some pieces of metadata can be seen as having 'more weight' than others.

This allows for a much more analog measurement of what a document is 'about'. A document can be 'about' many things, but to different extents. This brings us back to the way tags work in LibraryThing – many people can allocate different tags to the same book, and this allows a much more complex representation of 'aboutness'.

I don't think that this just applies to 'aboutness'. I believe other pieces of metadata could also benefit from an analogue approach – but I think I'm going to have to save this argument for another post.

The key thing here (for me) is that exploiting this linking and the network built using them is something that already exists – it is the web – and with it this brings a way of breaking out of our 'digital' approach to library data, that card or printed catalogues had to adopt by their very nature.

If every book in your catalogue had it's own URL – essentially it's own address on your web, you would have, in a single step, enabled anyone in the world to add metadata to the book – without making any changes to the record in your catalogue. I'd go further than this – but again that's going to need a post of its own – I hope I manage to get these written!

So, we have the means of enabling a much more sophisticated ('analogue') approach to metadata, and what is frustrating is that we have not yet realised this, and we still think 'digital data' is a 'pretty neat idea'.

A gathering place for UK Research

I’m the project director for EThOSNet – which is establishing a service, run by the British Library, to provide access to all UK PhD and Research Theses. The service itself is called EThOS (Electronic Theses Online Service).

Today, EThOS has gone into public beta – without fanfare, the service is now available, and can be found at http://ethos.bl.uk. The key parts of the service are:

  • A catalogue of the vast majority of UK Research Theses
  • The ability to download electronic versions where they exist
  • The ability to request an electronic version be created where it doesn’t already exist

I’m incredibly excited about this  – of all the projects I’ve been involved in, although not the biggest in terms of budget (I don’t think), it has the most potential to have an incredible impact of the availability of research. Until now, if you wanted to read a thesis you either had to request it via ILL, or take a trip to the holding university. Now you will now be able to obtain it online. To give some indication of the difference this can make, the most popular thesis from the British Library over the entire lifetime of the previous ‘Microfilm’ service was requested 58 times. The most popular electronic thesis at West Virginia University (a single US University) in the same period was downloaded over 37,000 times. If we can even achieve a relatively modest increase in downloads I’ll be happy – if we can hit tens of thousand then I’ll be delighted.

The project to setup EThOS has been jointly funded by JISC and RLUK, with contributions from the British Library, and a number of UK Universities and other partners, including my own, Imperial College London, which leads the project. The launch of the service is the culmination of several projects, including ‘Theses Alive!‘, ‘Electronic Theses‘, ‘DAEDALUS‘, ‘EThOS‘, and the current ‘EThOSNet‘.

With so much work done before and during the EThOSNet project, my own involvement (which started someway into the EThOSNet project, when I took over as Project Director from Clare Jenkins in autumn 2007), looks pretty modest, so thanks to all who have worked so hard to make EThOS possible, and get it live.

One of the biggest issues that has surfaced several times during the course of these projects, is the question of IPR (Intellectual Property Rights). EThOS is taking the bold, and necessary, step of working as an ‘opt-out’ service. This is based on a careful consideration of all the issues which has concluded:

  • The majority of authors wish to demonstrate the quality of their work.
  • Institutions wish to demonstrate the quality of their primary research

In order that authors can opt-out if they do not want their thesis to be made available via EThOS there is a robust take-down policy – available at EThOS Toolkit

As an author, you can also contact your University to let them know that you do not wish your thesis to be included in the EThOS service.

By making this opt-out and take-down approach as transparent as possible (including doing things like advertising it on this blog), we believe that authors have clear options they can exercise if they have any concerns about the service.

Finally, the derivation of the word Ethos (according to wikipedia) is quite interesting ™. There are many aspects of the word that felt relevant to the service – the idea of a ‘starting point’, and the idea that ‘ethos’ belongs to the audience both resonate with what EThOS is trying to do. However, for the title of the post I decided to draw on Michael Halloran’s assertion that "the most concrete meaning given for the term in the Greek lexicon is 'a habitual gathering place'." – which I believe is what EThOS will become to those looking for UK research dissertations.

Technorati Tags: ,

Developer Happiness Days

In February JISC is running dev8D – a 3/4 day event (9-13 Feb) aimed mainly at software developers in the UK HE space. There is also plenty of room for interested others at the event – as the website says:

Not an educational developer? Not a problem. One of the most exciting aspects of Developer Happiness Days is that it is gathering developers from a variety of sectors, both inside and outside the UK, to promote collaboration and creativity across different areas of research.

As well as teams from across the spectrum of higher education work, commercial developers are attending from a range of sectors. Providing an opportunity for public sector and commercial developers to meet and network is a key benefit of the event.

Not a developer? Again, welcome! The best code is created through collaboration with the end users and we want the people who will be testing the tools in their everyday work to have their voice heard too. And for the technophile tinkerers among you, there will be workshops to help bring everyone up to speed with the lingo.

I've been involved in the planning of the event (although this mainly consisted of commenting on other people's ideas!), but I hope you'll believe me when I say the final programme looks excellent.

The event is free to attend, and there is also a small amount of free accommodation available.

So, whether you are a developer / hacker / scriptkitty /

Technorati Tags:

JISC IE and eResearch Call briefing day

Today I’m at a JISC “town meeting” on the recent call for bids in the area of Information Environment and e-Research.  See also the live blog by Andy Powell at http://efoundations.typepad.com/livewire/2008/12/jisc-ie-and-eresearch-call-briefing-day.html

The meeting is essentially a briefing on the call, and a chance to ask questions, and hopefully get some idea of what might make a successful bid. There is lots of admin stuff first – who is eligible, what is expected from the bid – all of this is in the paperwork with the call, and you really need to go through this – this is all the essential stuff for having your bid considered, although a little bit dull!

A new aspect to this circular is that NERC, ESRC and EPSRC have expressed their willingness to work with projects – might be worth having a look at this (para 203).

Now into the meat of the briefing – first up, Matthew Dovey talking about the VRE programme

Phase 1 of the VRE programme looked at VLEs (Virtual Learning Environments), and wondered what a VRE (Virtual Research Environment) would look like. However, on both sides the idea of a ‘single box’ product for VLE or VRE has perhaps moved on – especially on the VRE side. Now the concept of a ‘VRE’ is more about a set of tools/services that support researchers in their work.

Phase 3 of the VRE programme is also about the ‘social’ side of the VRE – the ‘Virtual Research Community’ – building a community of practice about how you use VREs. There is an expectation that funded projects will actively engage with this community – and you should budget for this in the bid. Matthew making sure there is no doubt that this aspect is key to the programme.

Now Frederique van Till (Programme Manager for VRE Programme). She is saying the key for the programme is collaboration and community (as Matthew said). The VRC (Virtual Research Community) should bring together people/projects from phases 1, 2, and 3 (this one) of the programme. The focus is on technical interoperability nationally and institutionally (note the order of those!)

A diagram of the VRE Programme Phase 3 on the screen – not in the bid as it might be confusing (I wonder if it is available anywhere else?). Frederique introducing the idea of ‘Critical Friends’ – someone from outside the project, who is not linked to JISC, to help reflect on the project – these will be assigned to the project – so just be aware of this.

So – the VRE calls are:

B1. VRE Tools and Interoperability – 5 projects, up to £50k each, with projects expected to last 6-12 months (anything up to 2 years will be considered, but you’d have to show the budget would work!)

B2. VRE Frameworks – 4 projects, up to £150k each – up to 2 years

Some mention of Sakai, and ‘commercial’ solutions – look at the VRE phase 1 for starting points – and check eReSS wiki

B3. VRE National and Institutional Interoperability – 3-4 projects, up to £250k, expected to last 2 years

Projects should research both national and institutional interoperability (referred to here as ‘vertical’ and ‘horizontal’ respectively). Should look at the entire research lifecycle – Resource discovery, repositories, publishing and digital curation are strongly related to VREs – and this is where there is an overlap with the Information Environment projects – especially the A2 strand (see below).

This strand is looking for inter-disciplinary teams – researchers, librarians, programmers, usability experts, etc. etc.

Finally – make sure you build on lessons and work from phase 1 and phase 2, and look at the Strand B briefing note

Some quick questions:

Q: Stressed the importance of national interoperability – what about International interoperability?

A: International angles/collaboration encouraged, but bear in mind that funding is for UK

Q: Publication/Dissemination of research is mentioned in call – how much weight will be given to this?

A: It depends on the bids that come in, and the bids that come in under the IE strand. Looking for imaginative approaches. Interested in improving the process of dissemination (rather than the dissemination of the work from the projects per se)

Back after coffee – now the Information Environment strands of the call – although Rachel Bruce is noting that there will be further calls under the IE programme over the next few months.

Rachel acknowledging that the call has a focus on (institutional) repositories – noting some key projects from the past – e.g. FAIR, and outputs e.g. JORUM.

The current call reflects the move towards integration of digital repositories with other systems – e.g. Research Management systems, VLEs etc. Also network level services on the web – perhaps something that could be emphasised more in the call (says Rachel). Also the current call is looking at practical implementations based on pilots funded in earlier phases of the IE programme. There is an attempt to move focus from technical solutions to looking at the processes involved in research.

Note the DRIVER guidelines (European guidelines on repositories) – you may want to look at whether your project incorporates/exploits these.

As always with JISC calls, focus on ‘open’ – Open Source, Open Access etc. However, also a recognition that some data cannot be open.

Now Amber Thomas noting that the Open Education Resources call will be coming out later this week – over £5million across 3 strands for 12 month pilots, looking at the sustainable release of open learning content – looking at workflows and how practice is embedded. This funding is not for technical tools, or content creation. If you are interested in technical tools you need to be looking at the current call – especially the rapid innovation stuff.

Now Andy McGregor going to talk about:

A1. Automated metadata generation and text mining

Up to 5 projects over 18 months – £225k max per project – projects starting by 1st April 2009.

Metadata creation can be expensive and a lengthy process. With the increase of digital information production, manual creation processes struggle to keep up. Text-mining tools/techniques could help. Specifically the projects are deisgned to test if these tools can be put into practice, and if they can, what difference they make.

Evaluation is vital – need to look at the impact of the tools on the quality of the data and the cost benefit to the institution

N.B. The development of new automated meatadata generation and text mining tools are out of scope.

This is a call I’m quite interested in. My initial thoughts were around the EThOS service, and whether text mining could be applied to this body of text. However, from what Andy is saying, this would not fall into the criteria for this call directly. However, with the move to the submission of E-Theses at Imperial (all PhD theses submitted to Imperial now have to be submitted in electronic and print format), there seems to me to be potential here.

Also, we are buying more e-books – I wonder if there is some potential here?

Andy says bids must address:

  • Processes and content
  • Evaluation
    • Live service – compare the impact of the tools
    • Methodology – quantitative and qualitative – this needs to be made very clear in the bid – how you are going to evaluate impact is key to success – not enough to simply say ‘we will evaluate the effectiveness’
  • Document – need very clear documentation for< /li>
    • Rollout – how would a demonstrator be rolled out/scaled to go into production
    • Possibilities for other institutions – how would others take advantage
  • Hardware/Software
    • What tools are you going to exploit, what HW will you need, how are you going to knit tools together into integrated workflow

Rachel saying that you could be looking at a national level as well – which would bring EThOS back into scope? This needs more thought on my part – and discussion with others, especially current EThOSNet partners if this looks like a way to go.

If anyone else is interested in this area, and has some ideas on possible projects/collaborations please get in touch.

Now Neil Grindly talking about:

A2. Developing e-Infrastructure to support research disciplines

Projects in this strand have the largest amount of funding available – up to £1.35m per project (up to 2 projects, over 3 years)

In summary looking for:

  • Large ambitious prjects with a disciplinary focus
  • Involving collaboration across organisations and requiring varied skills
  • Questioning whether the current e-infrastructure is fit-for-purpose
  • Understanding emerging ways of doing research
  • Providing a number of related benefits to the sector
  • Taking advantage of and enhancing existing research and development
  • Providing sustainable and transferable benefits

Andy again on:

A3. Repositories start-up

These are matched funding projects – need 50% from the institution.

I had an interesting discussion with others at Imperial last week on establishing a Learning Object repository – I wonder if this would be something we could look at under this strand?

We are also trying to come to terms with research data

Andy says:

  • A range of content is in scope – learning materials, grey literature, data, multi-media etc. etc.
  • Must address a need
  • Could be institutional, departmental, disciplinary, etc.
  • It is valid to bid for a repository for a specific purpose, even if you already have one for other reasons – however, would be expected to deal with how repositories will interact
  • Projects will need to deal with technical implemenations, ploicy development, embedding and populating the repository
  • Digitisation or other content creation projects are out of scope
  • Previously funded startup projects are not eligible for further funding

There is lots of existing information in this area from previous startup projects, and the Repository Support Programme (RSP)

Make sure your bid considers:

  • Sustainability
  • Institutional commitment (i.e. you have real users from your community committed to the ideas of the repository)
  • How users needs will be addresses, and how they will be involved in the project
  • Compliance with the conditions set out in section 4 of the briefing document
  • Also look at the OER call if you are interested in making existing content available using existing tools

A4. Rapid Innovation

There are going to be 2 calls – so look out for second round of calls coming in the near future.

Repositories have lots of issues, but present many opportunities – these calls are for experimentation rather than sustainability – “Fail Forward Fast”

Projects that focus on a particular challenge or improvement for an established repository, on interoperability with institutional system or wider web environment, deal with widgets – are all welcome.

This call is about creative and innovative ideas – don’t be limited by the ideas outlined above.

Previous rapid innovation projects are:

  • Fedorazon
  • MR-CUTE
  • SNEEP

Key to these projects are:

  • Good communication – share outcomes/lessons learned
  • Dealing with established repositories/services rather than creating new ones
  • Engagement with the community
  • All staff must be in place for April 2009
  • Aspire to meet requirements in section 4 of the briefing document – but accept this may not always be possible/relevant

A5. Repository Enhancement

Open to all institutions in the UK (rather than HEFCE/HEFCW limits on other strands). Funding up to 22 projects, up to £350k (starting at £80k) over 2 years.

This strand is designed to focus on solving some tackling challenges to repositories:

  • Increasing content
  • Enhancing user experience – possibly with REF, could look at how information is produced
  • Preservation
  • Policy – how this can help embedding repositories – e.g. bringing in a mandate for deposit in the repository would be in scope (sounds interesting)

Look at partnerships with publishers, software suppliers etc. as well as others in education sector

Examples of work sought:

  • Embedding repository with other systems to make deposit easier
  • Interoperation with other institutional systems
  • Improving workflow for OER (Open Educational Resources)
  • Working with metadata
    • Scholarly communication metadata (e.g. ONIX for serials – can we exploit this?)
    • Taking advantage of metadata from elsewhere
    • Capturing metadata at source (e.g. scientific equipment generating material into repository – capture metadata directly)
  • Encouraging deposit

Need to include URL or screenshots for the repository you are going to enhance. Need to demonstrate clearly how the enhancements proposed could be exploited by other institutions, how you will build sustainability into what you do, how you are meeting user needs.

Here you should commit to meeting all relevant requirements in section 4 of the briefing document.

Previous enhancement projects are eligible for funding, but has to be a new project, not just an extension of the existing project.

Finally Neil again on:

A6. Preservation Exemplars

£900k available over up to 6 projects, 18 month duration, up to £225k per project.

Large range of scope here:

  • Dissemination activities
  • Emerging needs
  • e-journal archiving
  • Repositories and preservation
  • Record and asset management
  • Preservation of web resources
  • Preservation of e-learning material
  • Legal and economic policy and collaboration
  • Data curation
  • Digital object properties

Look to the Digital Curation Centre for support and advice.

3 stages of work:

  1. Needs and benefit analysis
  2. Pre-implementation planning and specification
  3. Implementation

Sorry – got distracted there, and didn’t really get this – but ideally JISC looking for projects that address all of these 3 stages, but recognise that the timescales may not always make this possible.

Questions and Answers

Places to find collaboration/air ideas:

OK – now lunch/networking. Questions (and answers I hope, although only Questions are on the agenda!) after lunch.

Before Q and A, a few bits and pieces:

  • Look at Project Management guidelines – important part of work
  • For advice you can look to UKOLN, JISC-CETIS, OSS Watch, DCC, Repositories Support Project
  • You may be interested in the JISC Developer Happiness Days – 9-13th February – keep an eye on http://dev8d.jiscinvolve.org – intended to help build an Innovation Community [declaration on interest, I was involved in the planning of this event in the form of c
    ommenting on and feeding into the structure and content of the event]

Q: Can the Rapid Innovation strand hope to get all projects in one go? Would it be better to spread out over a period of time?

A: There will be another round of Rapid Innovation funding. Potentially if JISC don’t receive 10 good proposals in the first round, they can review and carry funding over to future proposals

Q: Does the text-mining strand include (didn’t get the detail – a particular set of content?)

A: Will try to ensure there is funding for proposals across a variety of content

Q: Is partnership/collaboration compulsory?

A: No – although part of assessment. Recognise that large scale collaborations are more difficult to manage. Collaboration may apply to one part of project or all of it.

Q: Partnerships – some bids say ‘up to 2-4 partners’ – but some user groups are made up of more than this, and if you collaborate with user groups, you may end up with each constituent of a user group as a partner.

A: The size of partnerships mentioned are meant to be quite a strong steer – again, there is a size at which projects become very difficult to manage. You might want to consider how you relate to the various constituencies you interact with

Q: The bids are very structured into separate areas – VREs, VLEs, Institutional Repositories etc.Question of whether there is a reflection or steer as to where ‘repository’ people might be successful in terms of bids?

A: This call is a step towards bringing these various areas together – two strands ask explicitly for cross-domain teams. JISC organisation does reflect the structure to some extent, so understandable that you may be concerned about whether your area of work is relevant to the call. However, JISC sees these strands bringing these things together – so definite encouragement to bid  – “don’t be compartmentalised” (this is how the questioner put it, and was confirmed by Rachel from panel)

Q: We have been working on multi-media in e-portfolios. Would this kind of work fall into any of the call?

A: Yes – previous Repository/Startup Enhancement projects have tackled this kind of thing – look to these for examples. Definitely well within scope.

Q: In strand A2 “Projects teams must be cross-domain [I paraphrase] including academics and information/ICT professionals” – does this imply the institutional ICT/Library team?

A: Emphasis is cross-domain – so not completely necessary, but need to look at sustainability as well – so if this is something that institutional ICT/Library need to run longterm, wise to involve them. This certainly doesn’t exclude involving people with relevant skills from outside institutional services.

Q: A3. Would projects focusing on a specific kind of content be favoured over a project looking at all kinds of content?

A: No – but need to be clear on the institutional business need/business case

Q: Can same team work on more than one project?

A: No problem as long as very clear how they are separated – people allocated specifically and funded specifically. Need to avoid the projects being linked – i.e. it would be no good if you could not run one project if the second wasn’t funded

Q: [didn’t get this – something about cross-disciplinary research]

A: I think the answer was basically, this is fine – just need to be careful you are being realistic about what you can acheive within the project resources

Q: What relevant studies are coming up in the immediate future?

A: Sorry – I missed this, but apparently part of Frederique’s slides – which presumably will be made available.

Q: Something about using proprietary services or software, and what types of ‘open’ license are acceptable for software?

A: JISC is not against the use of proprietary services/software. However, be aware that you can get into problems if the plans of a commercial developer diverge from those of the project. If you are unclear what is appropriate, may be worth talking to relevant programme manager

Q: In strand A2, what about use of more lightweight/Web 2.0 approaches vs central/heavyweight infrastructure?

A: Definitely would consider both and a mixture of approaches across this spectrum. Doesn’t need to be ‘novel’ but does need to innovative. E-Infrastructure doesn’t imply big central services necessarily, so may be part of the picture. There was a followup question about tension between institutional needs and ‘collaborative’ research at a national level – the panel felt that any institution should be looking at collaborative (inter-institution) research.

And that is that – I’ve expressed an interest in a couple of the strands above (A1 and A3), so if you are also interested and might want to partner, drop me a line.

Technorati Tags:

A Suivre?

OK – a final post on Mashed Libraries, and then I’ll shut up. It has been really good to see blog posts start to drift in reflecting on the day, including:

These posts all (I’m glad to say) reinforce that the event was well worth doing. I’ve also got some more formal feedback via a survey (thank you Survey Monkey), and although there are certainly things to learn, and not everyone has had a chance to complete the survey yet, the overwhelming majority of respondents said they would attend a similar event in the future.

So – what comes next? On the day, there was enough interest in running something again next year. To be honest I’d prefer not to wait a whole 12 months, but sometimes these things take longer than you’d like. On the day we started to get ambitious with talk of 2 day events, unconferences and the like. I know OCLC are interested in doing something like the WorldCat Hackathon in Europe – perhaps we could think of a joint event?

Have a think about it, and let me know 🙂

In the meantime, here is a final way to savour the Mashed Libraries UK 2008 event. Thanks to Rob Styles and Dave Pattern for the photos, and rather appropriately the soundtrack is a track called ‘A Suivre’ by AlFa.