<?xml version="1.0" encoding="utf-8" ?><rss version="2.0"><channel><title>Bing: Databricks Multiple Use</title><link>http://www.bing.com:80/search?q=Databricks+Multiple+Use</link><description>Search results</description><image><url>http://www.bing.com:80/s/a/rsslogo.gif</url><title>Databricks Multiple Use</title><link>http://www.bing.com:80/search?q=Databricks+Multiple+Use</link></image><copyright>Copyright © 2026 Microsoft. All rights reserved. These XML results may not be used, reproduced or transmitted in any manner or for any purpose other than rendering Bing results within an RSS aggregator for your personal, non-commercial use. Any other use of these results requires express written permission from Microsoft Corporation. By accessing this web page or using these results in any manner whatsoever, you agree to be bound by the foregoing restrictions.</copyright><item><title>Is there a way to use parameters in Databricks in SQL with parameter ...</title><link>https://stackoverflow.com/questions/79035989/is-there-a-way-to-use-parameters-in-databricks-in-sql-with-parameter-marker-synt</link><description>Databricks demands the use of the IDENTIFIER () clause when using widgets to reference objects including tables, fields, etc., which is exactly what you're doing.</description><pubDate>Thu, 16 Apr 2026 22:25:00 GMT</pubDate></item><item><title>Databricks shows REDACTED on a hardcoded value - Stack Overflow</title><link>https://stackoverflow.com/questions/75753521/databricks-shows-redacted-on-a-hardcoded-value</link><description>It's not possible, Databricks just scans entire output for occurences of secret values and replaces them with " [REDACTED]". It is helpless if you transform the value. For example, like you tried already, you could insert spaces between characters and that would reveal the value. You can use a trick with an invisible character - for example Unicode invisible separator, which is encoded as ...</description><pubDate>Wed, 15 Apr 2026 01:41:00 GMT</pubDate></item><item><title>Printing secret value in Databricks - Stack Overflow</title><link>https://stackoverflow.com/questions/69925461/printing-secret-value-in-databricks</link><description>2 Building on @camo's answer, since you're looking to use the secret value outside Databricks, you can use the Databricks Python SDK to fetch the bytes representation of the secret value, then decode and print locally (or on any compute resource outside of Databricks).</description><pubDate>Fri, 17 Apr 2026 22:53:00 GMT</pubDate></item><item><title>how to get databricks job id at the run time - Stack Overflow</title><link>https://stackoverflow.com/questions/79659213/how-to-get-databricks-job-id-at-the-run-time</link><description>I am trying to get the job id and run id of a databricks job dynamically and keep it on in the table with below code run_id = self.spark.conf.get ("spark.databricks.job.runId", "no_ru...</description><pubDate>Thu, 16 Apr 2026 20:02:00 GMT</pubDate></item><item><title>Databricks shared access mode limitations - Stack Overflow</title><link>https://stackoverflow.com/questions/77216325/databricks-shared-access-mode-limitations</link><description>Databricks shared access mode limitations Ask Question Asked 2 years, 6 months ago Modified 2 years, 6 months ago</description><pubDate>Thu, 16 Apr 2026 00:43:00 GMT</pubDate></item><item><title>Databricks - Download a dbfs:/FileStore file to my Local Machine</title><link>https://stackoverflow.com/questions/66685638/databricks-download-a-dbfs-filestore-file-to-my-local-machine</link><description>Method3: Using third-party tool named DBFS Explorer DBFS Explorer was created as a quick way to upload and download files to the Databricks filesystem (DBFS). This will work with both AWS and Azure instances of Databricks. You will need to create a bearer token in the web interface in order to connect.</description><pubDate>Wed, 15 Apr 2026 22:20:00 GMT</pubDate></item><item><title>databricks: writing spark dataframe directly to excel</title><link>https://stackoverflow.com/questions/59107489/databricks-writing-spark-dataframe-directly-to-excel</link><description>Are there any method to write spark dataframe directly to xls/xlsx format ???? Most of the example in the web showing there is example for panda dataframes. but I would like to use spark datafr...</description><pubDate>Tue, 14 Apr 2026 05:24:00 GMT</pubDate></item><item><title>How to enable Databricks Connect in Visual Studio Code (Windows)?</title><link>https://stackoverflow.com/questions/77831431/how-to-enable-databricks-connect-in-visual-studio-code-windows</link><description>Enable Databricks Connect along with the Databricks Extension for Visual Studio Code to enhance your data processing capabilities. Databricks Connect allows you to seamlessly access remote Databricks clusters from your local environment, enabling you to execute notebooks and perform data queries as if you were directly connected to the cluster.</description><pubDate>Thu, 16 Apr 2026 00:14:00 GMT</pubDate></item><item><title>ValueError: cannot configure default credentials - Stack Overflow</title><link>https://stackoverflow.com/questions/78989067/valueerror-cannot-configure-default-credentials-visual-studio-and-databricks</link><description>I am trying to connect to Databricks with Visual Studio Code Databricks extension. Get following error: databricks-connect test In the extension everything looks fine I am logged into Databricks.</description><pubDate>Fri, 17 Apr 2026 11:11:00 GMT</pubDate></item><item><title>How to use python variable in SQL Query in Databricks?</title><link>https://stackoverflow.com/questions/72500067/how-to-use-python-variable-in-sql-query-in-databricks</link><description>I am trying to convert a SQL stored procedure to databricks notebook. In the stored procedure below 2 statements are to be implemented. Here the tables 1 and 2 are delta lake tables in databricks c...</description><pubDate>Fri, 17 Apr 2026 07:01:00 GMT</pubDate></item></channel></rss>