About 50 results
Open links in new tab
  1. Is there a way to use parameters in Databricks in SQL with parameter ...

    Sep 29, 2024 · Databricks demands the use of the IDENTIFIER () clause when using widgets to reference objects including tables, fields, etc., which is exactly what you're doing.

  2. Printing secret value in Databricks - Stack Overflow

    Nov 11, 2021 · 2 Building on @camo's answer, since you're looking to use the secret value outside Databricks, you can use the Databricks Python SDK to fetch the bytes representation of the secret …

  3. Databricks shows REDACTED on a hardcoded value - Stack Overflow

    Mar 16, 2023 · It's not possible, Databricks just scans entire output for occurences of secret values and replaces them with " [REDACTED]". It is helpless if you transform the value. For example, like you …

  4. databricks - How to create Storage Credential using Service Principal ...

    Sep 24, 2024 · An Azure Databricks access connector is a first-party Azure resource that lets you connect managed identities to an Azure Databricks account. You must have the Contributor role or …

  5. Databricks shared access mode limitations - Stack Overflow

    Oct 2, 2023 · Databricks shared access mode limitations Ask Question Asked 2 years, 6 months ago Modified 2 years, 6 months ago

  6. REST API to query Databricks table - Stack Overflow

    Jul 24, 2022 · Is databricks designed for such use cases or is a better approach to copy this table (gold layer) in an operational database such as azure sql db after the transformations are done in pyspark …

  7. How to use python variable in SQL Query in Databricks?

    Jun 4, 2022 · I am trying to convert a SQL stored procedure to databricks notebook. In the stored procedure below 2 statements are to be implemented. Here the tables 1 and 2 are delta lake tables …

  8. Databricks - Download a dbfs:/FileStore file to my Local Machine

    Method3: Using third-party tool named DBFS Explorer DBFS Explorer was created as a quick way to upload and download files to the Databricks filesystem (DBFS). This will work with both AWS and …

  9. Installing multiple libraries 'permanently' on Databricks' cluster ...

    Feb 28, 2024 · Installing multiple libraries 'permanently' on Databricks' cluster Asked 2 years, 1 month ago Modified 2 years, 1 month ago Viewed 5k times

  10. How to import own modules from repo on Databricks?

    Dec 7, 2022 · I have connected a Github repository to my Databricks workspace, and am trying to import a module that's in this repo into a notebook also within the repo. The structure is as such: …