Given a dataset node ID, runs processor-pre-external-files to download the latest curation export file for the given dataset to $INPUT_DIR
Also runs processor-pre-metadata to download the current metadata for the dataset to $INPUT_DIR
.
To build:
docker build -t pennsieve/ttl-sync-pre-processor .
On arm64 architectures:
docker build -f Dockerfile_arm64 -t pennsieve/ttl-sync-pre-processor .
To run tests:
go test ./...
To run integration test:
- Given a dataset you want to test with, create an integration for the dataset and this pre-processor. Get the integration id
- Copy
dev.env.example
todev.env
- In dev.env update SESSION_TOKEN with a valid token and INTEGRATION_ID with the id from the first step.
- Run
./run-integration-test.sh dev.env
If ENVIRONMENT=dev
is in dev.env
, then the processor will try to fetch the curation-export files from a test server
and you will have to make sure that there are files there to find.
If ENVIRONMENT=prod
is in `dev.env', the processor will use the real SPARC endpoint to find the curation-export files.