Replies: 1 comment 1 reply
-
So the package that gets pushed to Databricks doesn't get installed in the
|
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Is it possible to access package data files when running code on Databricks using the step launcher?
I've set up data files in our package like this: https://setuptools.pypa.io/en/latest/userguide/datafiles.html#subdirectory-for-data-files - using the "Subdirectory for Data Files" approach.
It's then possible to access them like this:
I guess this works because I have our package installed locally for development using
pip install -e .
. However I'm not sure the package gets installed like this by Dagster when it gets uploaded to Databricks? (using thelocal_pipeline_package_path
option).Is this approach to data files possible with the Dagster Databricks integration? It's a little tricky to debug remotely. I just want to make sure it's not a dead end before investing more time trying to get it to work.
Update with code:
The following asset code works when run without any IO manager: (spectacle is the module we pass to Dagster as
module_name
)When I add our Databricks step launcher IO manager the code errors on the
files(...)
line. This is the error:Beta Was this translation helpful? Give feedback.
All reactions