A collection of Python/SQL scripts for data cleanup, management, and others related to the Smithsonian's ArchivesSpace implementation.
This script takes a CSV of resources and archival objects from every repository with "Missing Title" titles in note lists and removes the title from the metadata, then posts the update to ArchivesSpace
- ArchivesSnake
- ArchivesSpace username, password, API URL in a secrets.py file
- logs directory for storing local log files
- test_data/missingtitles_testdata.py file, with the following:
test_object_metadata = {ArchivesSpace resource or archival object metadata}
for testing. Can get this from your API by using aclient.get
request for a resource or archival object that has a "Missing Title" in one of its notes with a list.test_notes = [ArchivesSpace resource or archival object notes list]
for testing. Can get this from your API using aclient.get
request for a resource or archival object that has a "Missing Title" in one of its notes with a list and taking all the data found in"notes" = [list of notes]
- test_data/MissingTitles_BeGone.csv - a csv file containing the URIs of the objects that have "Missing Title" in their
notes. URIs should be in the 4th spot (
row[3]
)
Unittests for missingtitles_tests.py
This script collects all users from ArchivesSpace, parses their usernames to separate any starting with 'z-' and ending with '-expired-' into just the text in-between, then updates the username in ArchivesSpace with the new username
- ArchivesSnake
- ArchivesSpace username, password, API URL
- logs directory for storing local log files
- test_data/znames_testdata.py file, with
viewer_user = {ArchivesSpace viewer user metadata}
for testing. Can get this from your API by getting aclient.get
request for the `viewer' user in your ArchivesSpace instance.
Unittests for znames_tests.py
Not every script requires every package as listed in the requirements.txt file. If you need to use a script, check the import statements at the top to see which specific packages are needed.
- ArchivesSnake - Library used for interacting with the ArchivesSpace API
- loguru - Library used for generating log files
- Download the repository via cloning to your local IDE or using GitHub's Code button and Download as ZIP
- Run
pip install requirements.txt
or check the import statements for the script you want to run and install those packages - Create a secrets.py file with the following information:
- An ArchivesSpace admin username (as_un = ""), password (as_pw = "")
- The URLs to your ArchivesSpace staging (as_api_stag = "") and production (as_api = "") API instances
- Your ArchivesSpace's staging database credentials, including username (as_dbstag_un = ""), password (as_dbstag_pw = ""), hostname (as_dbstag_host = ""), database name (as_dbstag_database = ""), and port (as_dbstag_port = "")
- Create a logs folder in your project's local directory for storing log files
- Run the script as
python3 <name_of_script.py>
Each script has its own parameters, most not requiring any arguments to run. However, you will want to take time to adjust the script to meet your own needs. For instance, you may want to set up a 'data' and/or 'reports' folder in your code's directory to store exported CSV's, Excel spreadsheets, or any other outputs that are generated from the script. See the Overview section for more info on what each script does.
- Select which script you would like to run
- Run the script with the following command for python scripts:
python3 <name_of_script.py>
- If there are arguments, make sure to fill out those arguments after the python script name. Most scripts just need the information listed in secrets.py file created in the installation step above.
- If the script is not a python script, but an SQL statement, you can either download the SQL file or copy the code to your local SQL developer environment and run it there.
- Corey Schmidt - IT Specialist at the Smithsonian Institution
- ArchivesSpace community