About 50 results
Open links in new tab
  1. Is there a way to use parameters in Databricks in SQL with parameter ...

    Sep 29, 2024 · Databricks demands the use of the IDENTIFIER () clause when using widgets to reference objects including tables, fields, etc., which is exactly what you're doing.

  2. Printing secret value in Databricks - Stack Overflow

    Nov 11, 2021 · 2 Building on @camo's answer, since you're looking to use the secret value outside Databricks, you can use the Databricks Python SDK to fetch the bytes representation of the secret …

  3. Databricks shared access mode limitations - Stack Overflow

    Oct 2, 2023 · You're correct about listed limitations. But when you're using Unity Catalog, especially with shared clusters, you need to think a bit differently than before. UC + shared clusters provide very …

  4. Databricks - Download a dbfs:/FileStore file to my Local Machine

    Method3: Using third-party tool named DBFS Explorer DBFS Explorer was created as a quick way to upload and download files to the Databricks filesystem (DBFS). This will work with both AWS and …

  5. databricks: writing spark dataframe directly to excel

    Nov 29, 2019 · Are there any method to write spark dataframe directly to xls/xlsx format ???? Most of the example in the web showing there is example for panda dataframes. but I would like to use spark …

  6. How to use python variable in SQL Query in Databricks?

    Jun 4, 2022 · I am trying to convert a SQL stored procedure to databricks notebook. In the stored procedure below 2 statements are to be implemented. Here the tables 1 and 2 are delta lake tables …

  7. Installing multiple libraries 'permanently' on Databricks' cluster ...

    Feb 28, 2024 · Installing multiple libraries 'permanently' on Databricks' cluster Asked 2 years, 1 month ago Modified 2 years, 1 month ago Viewed 5k times

  8. Databricks change default catalog - Stack Overflow

    Jun 29, 2023 · But Databricks recommends keeping the default catalog as hive_metastore, because changing the default catalog can break existing data operations that depend on it.

  9. install python packages using init scripts in a databricks cluster

    Jun 22, 2020 · I have installed the databricks cli tool by running the following command pip install databricks-cli using the appropriate version of pip for your Python installation. If you are using Python …

  10. Convert string to date in databricks SQL - Stack Overflow

    Jul 9, 2021 · Use Databricks Datetime Patterns. According to SparkSQL documentation on the Databricks website, you can use datetime patterns specific to Databricks to convert to and from date …