-
Notifications
You must be signed in to change notification settings - Fork 19
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add forge methods to acces/query datasets from external sources #367
Draft
crisely09
wants to merge
50
commits into
master
Choose a base branch
from
dev-datasets
base: master
Could not load branches
Branch not found: {{ refName }}
Loading
Could not load tags
Nothing to show
Loading
Are you sure you want to change the base?
Some commits from the old base branch may be removed from the timeline,
and old review comments may become outdated.
Conversation
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
* load local configuration when configuration tests * keep commons imports together * fix duplicate keyword argument issue * use context path from default config in test_to_resource test * rm extra store_config.pop * refactor store config
* pass view when sparql, elastic call, todo search * rm unless constants from store * turn view to endpoint * endpoint to view param * rename param * rename param2 * keyword param from forge call * missing underscore * git status * make endpoint refac * edit querying notebook to showcase feature, todo set up better view * refac notebook edit * change view creation mapping * check filters not provided as keyword arg * fix querying notebook, retrieve using resource url * test make endpoint function * use *filters for the store interface and implementations
* added timeouts to every requests. call * centralise default request timeout * rm import * use constant in file get
…ething it raises an error
* change signatures to allow for boolean change_schema on update * change_schema implemented for single resource update * refactor batch actions * lint * began batch update schema * change schema many * progress * rm / join * rm useless change * fix * change schema to update schema * update instead of change * example notebook for schema method, todo update * notebook example with update * lint * improve notebook * fix one test * keep unconstrained schema only for update endpoint else _ * same url building in one and many * add timeout * schema id optional in update * rename local parse * rename keep_unconstrained to use_unconstrained_id * rm extra docstring param
* change signatures to allow for boolean change_schema on update * change_schema implemented for single resource update * refactor batch actions * lint * began batch update schema * change schema many * progress * rm / join * rm useless change * fix * change schema to update schema * update instead of change * example notebook for schema method, todo update * notebook example with update * lint * improve notebook * fix one test * keep unconstrained schema only for update endpoint else _ * same url building in one and many * add timeout * schema id optional in update * rename local parse * rm second request for metadata * add query param annotate only if retrieve source is true * retrieval error if issue creating resource from response * rm cross bucket check with source * add todo * separate metadata call if cross bucket and source * refac * fixes and notebook update * check deployment endpoint self, may need to be checking multiple values * revert to self.endpoint * updated notebook to show retrieval * fix response name * add query param annotate only if retrieve source is true * separate metadata call if cross bucket and source * rename keep_unconstrained to use_unconstrained_id * rm extra docstring * better comments * clarify comment * comment fix * improve markdown
* re-do metadata fetch until endpoint is fixed * better notebook * rename variables * code style * fix replace in comments * update comments
…False (#382) * return resource as_json optionally when forge.elastic * as_resource instead of as_json, default True * skeleton to enable building resources from different values in the es response payload * example of forge.elastic as_resource = False in getting started Querying notebook
* Set jsonld context version to 1.1 * this enables non IRI-delimiting character not present in rdflib.plugins.shared.jsonld.context.URI_GEN_DELIMS (e.g '_' in "NCBITaxon": "http://purl.obolibrary.org/obo/NCBITaxon_") to be used when defining jsonld context prefix
Currently the pySHACL throws a ReportableRuntimeError("Evaluation path too deep!\n{}".format(path_str)) exception when evaluating a shape if the length of its transitive closure of the sh:node property is bigger or equal to 30. Give a node shape, this PR addresses this by: * Fixes pyshacl deep nodeshape path eval error by first recursively collecting all the property shapes directly defined (through sh:property) by the node shape or inherited from the node shape parents then link those collected property shapes to the node shape through sh:property and finally remove the node shape <-> parent shape relatioinships * Aligned the expected data model for a shacl shape between the RDF StoreService and DirectoryService:
* Fixed coupled of issues: * forge.prefixes() was raising pandas.Dataframe error "if using all scalar values, you must pass an index" * fixed forge.types() to properly collect types from rdfsercice.class_to_shape * fixed forge.template() when using an rdf model based on a store service: Unable to generate template:'tuple' object has no attribute 'traverse'
* Added support for inference when using the RdfModel: * support for importing ontologies from schemas using owl:imports * use forge.validate(resource, inference="inference_value", type_='AType') with inference_value as in https://github.com/RDFLib/pySHACL/blob/v0.25.0/pyshacl/validate.py#L81. inference_value="rdfs" seems to be enough to extend the resource with the transitive closures of type subClassOf and/or property subPropertyOf relations as per the RDFS entailment rules (https://www.w3.org/TR/rdf-mt/). * Validation now fails when a type not in the resource is provided as value of the type_ argument unless inference is enabled (with inference='rdfs' for example) and the resource type is a subClassOf of type_
…cursive_resolve when resolving a str jsonld context (#402)
* Add alternateName to agent resolver * Added property also to the agent-to-resource mapping
resolve with target species returns only species and strain returns strains
…able (#406) * Add method when initializing forge to export environment variable * Remove addition in setup.py and try os.environ instead of os.system
…-file-content-length` to header (#403) * Remove nexus_sdk from nexus store when uploading files * Change content length header to be
* split file get call and property access * split prepare download one * lint --------- Co-authored-by: Leonardo Cristella <[email protected]>
* rm sdk usage from bluebrainexus store file * rm sdk usage from utils * rm sdk usage from service * rm nexussdk from test * lint * fix patch * change usage of project_fetch function for successful patching in tests * rename module of sdk methods * remove leftover image from store * missing file * restore config * remove s from file name * missing the file, again --------- Co-authored-by: Cristina E. González-Espinoza <[email protected]>
* Added batched retrieval
* Check for schema id to use schema endpoint * Add example in unit test * Add missing parameter inside _register_one. Add notebook example of schema handling * check if resource_id is given
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
This is not complete, there are inconsistencies to be fixed, but I didn't manage to do it on time.
Basically, IT DOESN'T WORK YET.
I leave it open for now.