KISS "Keep it simple" - Solutions for the underfunded (Jody)
Lots of institutions out there have been or are creating digital content, and have no preservation plans or methods in place.
Many of them have no programmers or even metadata librarians. What options do they have for having a prayer that all their
hard work and precious digital content may survive?
Internet archive -- accept bare minimum of metadata -- title most important.
Some file format dependent metadata required. Will retain bit level preservation.
Contact state library -- see if they have a ingest.
In CA (Joan Starr) there's a LHGRP local history digital -?- program, which provides
grants to help local tiny museums digitize and get content into Online Archive of Cal.
Changed approach to Pitch-an-idea this year.
Provides consultative support, keeps archival content, not just jpegs.
If someone already has content in other standards -- do they have to rework it for OAC? unknown
Naomi ? worked with NSDL to collect very small collections -- used OAI -- but difficult, because small
institutions didn't have that skill set.
Bob Horton at Minn. Digital library -- working with small institutions -- one-off way, different with
each grant. Want to consolidate it -- storing some at the archive.
Working with Internet Archive to streamline uploads.
Eric Celeste just did a report for them. Access to report?
LOC digital stewardship allicances -- "digital preservation in a box" -- guidelines -- follow--on of NDIIPP
Outline of a workshop.
NEW -- still ironing it out. Now have public link.
National digital stewardship alliance membership is free on institutional basis, simply apply.
What is preserved??
CSV ok, not necessary to have XML
File structure can be echoed in directory structure.
Insufficient time to catalog -- how can we make material accessible?
Are people comfortable with user-contributed metadata?
What about Google image labeler?
Brings up image, putting in keywords, only stick if verified by another person.
Museums are using it -- like STEVE keyword -- but has verification mechanisms.
Museums don't tend to use keywords.
If Google could open this for more cultural heritage institutions, it would be great.
Kam Woods says anyone can contribute content, and applied keywords can be exported back out.
SPKARD.com (?) Some company that presented at the Soc. for Cal. Archivists in the spring --
Makes money by organizing volunteers online -- recruits companies to participate.
Many are not even getting their content on the web -- people like to portalize, but that's awful.
Need Google to be able to index content intelligently.
Need indexing and access cheap fast and free -- via Google
Need URLs that match the organization of data on the back end, if designed into system early on.
URLs that make sense, can strip off segments of URL tree.
The verification available via google image is great-- provides check and balance.
Locally might need a second part of the database for use-contributed keywords; desire to keep separate from librarian/archivist created metadata.
Lots of people want to contribute -- need ways to auto-curate it.
Need practical, step by step way to de-silo content, bring it up to the level of the web.
Need simple methods to make content accessible by google bots
Provide institutions with a script to collect their content, stick it in web directories, wrap in HTML
Just getting content into the web is a really high bar for many people, may not even have access to web hosting.
Could reach out to business world in local area --
2cella -- safe deposit box -- starting to work with states to offer as a hosted service.
Omeka.net now for hosted services
Still have to pay. Money is always a barrier.
Internet archive costs, but is free for contributors. Devoted to store content for those who can't host it for free.
IA Contributors still need cloud provider -- will cost for that.
Amazon S3 instance -- portals will set that up for you.
NSDL is free in a way -- offered 100 TB but they didn't write back; what do you do with huge amounts of data?
IA will take it, but batch processing requires programmer.
IA staffed to maintain integrity of the data they're given, not to assist in ingest.
How do you balance the huge users against the small ones? IA -- mostly one-offs, no competition.
Lots of small museums use Past Perfect -- a little virtual exhibit site, creating HTML
Digitized according to whatever non-standard -- no hi-res images made accessible.
Many want to control access to hi-res images, can they do that via IA?
How to get over the policy barrier for hi-res -- CAN make it dark in internet archive -- can't make it gray at this point. Don't like to make it dark. Must have copyright status to do so.
Brewster Kahle might consider the gray aspect if that's the only way to keep the content safe.
For higher ed institutions, Write it into the agreements -- must allow digital preservation of hi-res images and archival content.
Front-end to IA? not much developed. Would be helpful if useful front-end. providing access service
search and browse features, use IA as server of hi and medium-resolution images.
IA has a difficult serials issue
openlibrary.org displays IA books
User interface needs improvement. open-source development group of volunteers?
need better public API.
(Deborah Holmes-Wong) There are some state and regional consortiums. Higher ed institutions will help if involved on the front end, to control metadata and standards,
provide standards and training, even with volunteers. most projects via faculty members. (USC)
However, this sometimes makes the barrier too high for them to sustain, manage, or even attempt.
We want to empower them, but provide them with something they can sustain.
We must meet people on their own ground.