Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Load data into databricks without external staging and auth. #2166

Open
rudolfix opened this issue Dec 19, 2024 · 0 comments
Open

Load data into databricks without external staging and auth. #2166

rudolfix opened this issue Dec 19, 2024 · 0 comments
Assignees
Labels
support This issue is monitored by Solution Engineer

Comments

@rudolfix
Copy link
Collaborator

It seems that python sdk for databricks allows to upload files.

  1. Research if it is possible to load files into tables like we do into BigQuery: when a local file may be copied into a table without any stage
  2. If that does not work, research how to use Volumes on databricks to copy files there and use COPY INTO to move them into table.
  3. If authentication is not configured, enable default credentials (ie. if present on serverless compute). You can take a look how CredentialsWithDefault is used (most implementations check if default credentials are present in def on_partial(self) -> None: but in your case you should to it in on_resolve when all fields holding credentials are empty)

Ideal scenario. when running in a Notebook, is that we can load a source (ie rest_api) without any additional configuration, staging or authorization - like we are able to do with duckdb

@rudolfix rudolfix moved this from Todo to In Progress in dlt core library Dec 19, 2024
@rudolfix rudolfix added the support This issue is monitored by Solution Engineer label Dec 19, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
support This issue is monitored by Solution Engineer
Projects
Status: In Progress
Development

No branches or pull requests

2 participants