Hi,
I was talking to Brian Thomas about this group at the Best Practices Exchange Conference recently and I'd love to resurrect it. As a start I have a Preservica problem. I want to bulk upload essentially a hard drive's amount of files and then link the file to its existing ArchivesSpace archival object. The problem is automating that. Preservica Support gave me information about their OPEX ingest function but I need to create the OPEX files and insert them into the right directory level while also querying the ArchivesSpace API via the component identifier because that's the only link I have to the ArchivesSpace object.
My Questions are:
1. How to get the Archival Object ID via the ArchivesSpace API with only the component identifier field as an access point
2. The best way to generate the OPEX metadata. I've been making it in a spreadsheet then converting it to XML
3. Inserting the right metadata file in the right directory for ingest.
Attached are the python scripts I have so far:
opexTransform.py reads the spreadsheet and converts it to XML by cheating and just making a text file that happens to be XML.
patternDir.py reads the original file directory and copies the directory structure and only the preservation file formats, ignoring any access or vendor created admin files.
I have nothing for the ArchivesSpace part. I'm more comfortable with the ArchivesSpace database but my new institution hosts their instance so I don't have access to it.
------------------------------
Corinne Chatnik
Union College, Shaffer Library
Schenectady, NY
chatnikc@union.edu------------------------------