<?xml version="1.0" encoding="utf-8" ?><rss version="2.0"><channel><title>Bing: Databricks Square Logo</title><link>http://www.bing.com:80/search?q=Databricks+Square+Logo</link><description>Search results</description><image><url>http://www.bing.com:80/s/a/rsslogo.gif</url><title>Databricks Square Logo</title><link>http://www.bing.com:80/search?q=Databricks+Square+Logo</link></image><copyright>Copyright © 2026 Microsoft. All rights reserved. These XML results may not be used, reproduced or transmitted in any manner or for any purpose other than rendering Bing results within an RSS aggregator for your personal, non-commercial use. Any other use of these results requires express written permission from Microsoft Corporation. By accessing this web page or using these results in any manner whatsoever, you agree to be bound by the foregoing restrictions.</copyright><item><title>Is there a way to use parameters in Databricks in SQL with parameter ...</title><link>https://stackoverflow.com/questions/79035989/is-there-a-way-to-use-parameters-in-databricks-in-sql-with-parameter-marker-synt</link><description>Databricks demands the use of the IDENTIFIER () clause when using widgets to reference objects including tables, fields, etc., which is exactly what you're doing.</description><pubDate>Thu, 02 Apr 2026 15:25:00 GMT</pubDate></item><item><title>Databricks shows REDACTED on a hardcoded value - Stack Overflow</title><link>https://stackoverflow.com/questions/75753521/databricks-shows-redacted-on-a-hardcoded-value</link><description>It's not possible, Databricks just scans entire output for occurences of secret values and replaces them with " [REDACTED]". It is helpless if you transform the value. For example, like you tried already, you could insert spaces between characters and that would reveal the value. You can use a trick with an invisible character - for example Unicode invisible separator, which is encoded as ...</description><pubDate>Tue, 31 Mar 2026 02:28:00 GMT</pubDate></item><item><title>Printing secret value in Databricks - Stack Overflow</title><link>https://stackoverflow.com/questions/69925461/printing-secret-value-in-databricks</link><description>2 Building on @camo's answer, since you're looking to use the secret value outside Databricks, you can use the Databricks Python SDK to fetch the bytes representation of the secret value, then decode and print locally (or on any compute resource outside of Databricks).</description><pubDate>Sat, 04 Apr 2026 07:09:00 GMT</pubDate></item><item><title>Where does databricks store the managed tables? - Stack Overflow</title><link>https://stackoverflow.com/questions/79164012/where-does-databricks-store-the-managed-tables</link><description>Answering your two sub questions individually below: Does this mean that databricks is storing tables in the default Storage Account created during the creation of Databricks workspace ? Yes. It stores the tables at the default location that is /user/hive/warehouse location. If the answer to above question is Yes, then is it a good practice to store tables here or should we store it in a ...</description><pubDate>Thu, 02 Apr 2026 17:06:00 GMT</pubDate></item><item><title>Newest 'databricks' Questions - Stack Overflow</title><link>https://stackoverflow.com/questions/tagged/databricks?tab=Newest</link><description>Use this tag for questions specific to Databricks Lakehouse Platform, including, but not limited to Databricks file system, REST APIs, Databricks Spark SQL extensions and orchestrating tools.</description><pubDate>Fri, 03 Apr 2026 17:19:00 GMT</pubDate></item><item><title>How to get usage statistics from Databricks or SQL Databricks?</title><link>https://stackoverflow.com/questions/75442193/how-to-get-usage-statistics-from-databricks-or-sql-databricks</link><description>Yes, there are several ways to get usage statistics from Databricks: Databricks UI: The Databricks UI provides information on the usage of tables, notebooks, and jobs.</description><pubDate>Sat, 04 Apr 2026 05:07:00 GMT</pubDate></item><item><title>Databricks api list all jobs from workspace - Stack Overflow</title><link>https://stackoverflow.com/questions/78758487/databricks-api-list-all-jobs-from-workspace</link><description>I am trying to get all job data from my Databricks. Basically, I need to put all job data into a DataFrame. There are more than 3000 jobs, so need to use the page_token to traverse all pages. Here ...</description><pubDate>Fri, 03 Apr 2026 04:40:00 GMT</pubDate></item><item><title>azure - Databricks - Read CSV file from folder - Stack Overflow</title><link>https://stackoverflow.com/questions/74153955/databricks-read-csv-file-from-folder</link><description>Databricks - Read CSV file from folder Asked 3 years, 5 months ago Modified 1 year, 11 months ago Viewed 13k times</description><pubDate>Thu, 02 Apr 2026 21:23:00 GMT</pubDate></item><item><title>How to use python variable in SQL Query in Databricks?</title><link>https://stackoverflow.com/questions/72500067/how-to-use-python-variable-in-sql-query-in-databricks</link><description>I am trying to convert a SQL stored procedure to databricks notebook. In the stored procedure below 2 statements are to be implemented. Here the tables 1 and 2 are delta lake tables in databricks c...</description><pubDate>Thu, 02 Apr 2026 16:01:00 GMT</pubDate></item><item><title>Databricks - Download a dbfs:/FileStore file to my Local Machine</title><link>https://stackoverflow.com/questions/66685638/databricks-download-a-dbfs-filestore-file-to-my-local-machine</link><description>Method3: Using third-party tool named DBFS Explorer DBFS Explorer was created as a quick way to upload and download files to the Databricks filesystem (DBFS). This will work with both AWS and Azure instances of Databricks. You will need to create a bearer token in the web interface in order to connect.</description><pubDate>Wed, 01 Apr 2026 20:20:00 GMT</pubDate></item></channel></rss>