KISS "Keep it simple" - Solutions for the underfunded (Jody)

From CURATEcamp
Jump to: navigation, search


Lots of institutions out there have been or are creating digital content, and have no preservation plans or methods in place.

Many of them have no programmers or even metadata librarians. What options do they have for having a prayer that all their

hard work and precious digital content may survive?

(Vickey Brasseur)

Internet archive -- accept bare minimum of metadata -- title most important.

Some file format dependent metadata required. Will retain bit level preservation.

Soliciting content.

Contact state library -- see if they have a ingest.

In CA (Joan Starr) there's a LHGRP local history digital -?- program, which provides

grants to help local tiny museums digitize and get content into Online Archive of Cal.

Changed approach to Pitch-an-idea this year.

Provides consultative support, keeps archival content, not just jpegs.

If someone already has content in other standards -- do they have to rework it for OAC? unknown

Naomi ? worked with NSDL to collect very small collections -- used OAI -- but difficult, because small

institutions didn't have that skill set.

Bob Horton at Minn. Digital library -- working with small institutions -- one-off way, different with

each grant. Want to consolidate it -- storing some at the archive.

Working with Internet Archive to streamline uploads.

Eric Celeste just did a report for them. Access to report?

LOC digital stewardship allicances -- "digital preservation in a box" -- guidelines -- follow--on of NDIIPP

Outline of a workshop.

NEW -- still ironing it out. Now have public link.

National digital stewardship alliance membership is free on institutional basis, simply apply.

What is preserved??

CSV ok, not necessary to have XML

File structure can be echoed in directory structure.

Insufficient time to catalog -- how can we make material accessible?

Are people comfortable with user-contributed metadata?

What about Google image labeler?


Brings up image, putting in keywords, only stick if verified by another person.

Museums are using it -- like STEVE keyword -- but has verification mechanisms.

Museums don't tend to use keywords.

If Google could open this for more cultural heritage institutions, it would be great.

Kam Woods says anyone can contribute content, and applied keywords can be exported back out. (?) Some company that presented at the Soc. for Cal. Archivists in the spring --

Makes money by organizing volunteers online -- recruits companies to participate.

Many are not even getting their content on the web -- people like to portalize, but that's awful.

Need Google to be able to index content intelligently.

Need indexing and access cheap fast and free -- via Google

Need URLs that match the organization of data on the back end, if designed into system early on.

URLs that make sense, can strip off segments of URL tree.

The verification available via google image is great-- provides check and balance.

Locally might need a second part of the database for use-contributed keywords; desire to keep separate from librarian/archivist created metadata.

Lots of people want to contribute -- need ways to auto-curate it.

Need practical, step by step way to de-silo content, bring it up to the level of the web.

Need simple methods to make content accessible by google bots

Provide institutions with a script to collect their content, stick it in web directories, wrap in HTML

Just getting content into the web is a really high bar for many people, may not even have access to web hosting.

Could reach out to business world in local area --

2cella -- safe deposit box -- starting to work with states to offer as a hosted service. now for hosted services

Still have to pay. Money is always a barrier.

Internet archive costs, but is free for contributors. Devoted to store content for those who can't host it for free.

IA Contributors still need cloud provider -- will cost for that.

Amazon S3 instance -- portals will set that up for you.

NSDL is free in a way -- offered 100 TB but they didn't write back; what do you do with huge amounts of data?

IA will take it, but batch processing requires programmer.

IA staffed to maintain integrity of the data they're given, not to assist in ingest.

How do you balance the huge users against the small ones? IA -- mostly one-offs, no competition.

Lots of small museums use Past Perfect -- a little virtual exhibit site, creating HTML

Digitized according to whatever non-standard -- no hi-res images made accessible.

Many want to control access to hi-res images, can they do that via IA?

How to get over the policy barrier for hi-res -- CAN make it dark in internet archive -- can't make it gray at this point. Don't like to make it dark. Must have copyright status to do so.

Brewster Kahle might consider the gray aspect if that's the only way to keep the content safe.

For higher ed institutions, Write it into the agreements -- must allow digital preservation of hi-res images and archival content.

Front-end to IA? not much developed. Would be helpful if useful front-end. providing access service

search and browse features, use IA as server of hi and medium-resolution images.

IA has a difficult serials issue displays IA books

User interface needs improvement. open-source development group of volunteers?

need better public API.

(Deborah Holmes-Wong) There are some state and regional consortiums. Higher ed institutions will help if involved on the front end, to control metadata and standards,

provide standards and training, even with volunteers. most projects via faculty members. (USC)

However, this sometimes makes the barrier too high for them to sustain, manage, or even attempt.

We want to empower them, but provide them with something they can sustain.

We must meet people on their own ground.