-
Notifications
You must be signed in to change notification settings - Fork 4
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Filtering out embargoed data #18
Comments
This Cypher pattern documents a standard pattern of relationships from a single individual MATCH (ds:DataSet) where ds.production
WITH ds
MATCH lp=(ds)-[:has_license]-(l:License)
WITH ds, lp
OPTIONAL MATCH pp=(ds)-[:has_reference]->(p:pub)
WITH ds, lp, pp
MATCH ip=(ds)<-[:has_source]-(i:Individual)<-[:depicts]-(ch:Individual)-[:in_register_with]->(tc:Individual)-[:depicts]-(t)
WITH ds, lp, pp, ip, i
MATCH icp=(i)-[:INSTANCEOF|Related|hasDbXref]->(c)
RETURN lp,pp, ip, icp limit 2 TBA - (i)-[]-(i) where bother are in production In addition - all classes and relationships between them should be loaded. |
Alternative approach - delete entities under embargo after full release
|
If there is no production flag = true, delete. |
Flag for broken images that should be filtered out. |
2 flags:
We want to be able to run a branch of the pipeline in staging mode. |
blocking on plugin level. |
- Added SHACL pipeline #3 - Found first solution for SPARQL based embargo (VirtualFlyBrain/neo4j2owl#18)
|
Manual tests: Exploring a non-embargoed channel from the Aso2014 dataset: Exploring an embargoes channel from the TrumanWood2018 dataset: Exploring all mentions of the production property: @dosumis this is now ready for your review. |
We need to add a filtering step so that datasets can be embargoed. In pipeline 1, this step is at the OWL export stage. In p2 it should also be between KB and the integration layer.
The step is essential to having a complete, functioning pipeline.
Filtering works by matching a graph pattern, starting from DataSet. Note, graph patterns specifying filters are likely to be close or identical to the ones we need to schema validation. We can base specification on the filtering code used in owl generation.
The text was updated successfully, but these errors were encountered: