Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
This one changes how data validation is handled in cascade. It introduces schemas in datasets.
Now pipeline blocks can explicitly define input schemas as pydantic models (optional dependency for this feature). If schema is defined in new
SchemaModifier
then it will implicitly insertValidationWrapper
before itself in the pipeline. When usingself._dataset
inside its own__getitem__
it will wrap itself in the validator which will perform validation of the input.For example we have a dataset of annotated images:
Base class for the pipeline will be:
Let's define a dummy modifier:
Let's say we have a source of images and
hope that it will maintain our schema. The following
is the regular way we use modifier, but under the hood
it will automatically check that the output of the
datasource is AnnotImage