adding s3 bucket parsing for model paths #81
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
When using
import-model
to import from an exported open source project to Databricks, I was receiving the following error:This is due to the source field in
model.json
being an S3 bucket that contains "artifacts" (i.e.s3://test-analytics-us-east-1-mlflow-artifacts
) which results in the pattern "artifacts" occurring twice, resulting in:dbfs:/databricks/mlflow-tracking/2485056402562980/6ddc7ac453ea44abb2f740628fe3b423/artifacts/45/007930631053402f957be01de073fc3e/artifacts/model
instead of:
dbfs:/databricks/mlflow-tracking/2485056402562980/6ddc7ac453ea44abb2f740628fe3b423/artifacts/model
.I've added a check for source S3 URLs and parses the path so that the bucket name will no longer interfere with the creation of the
model_path
variable.