About 50 results
Open links in new tab
  1. Printing secret value in Databricks - Stack Overflow

    Nov 11, 2021 · 2 Building on @camo's answer, since you're looking to use the secret value outside Databricks, you can use the Databricks Python SDK to fetch the bytes representation of the secret …

  2. Databricks shows REDACTED on a hardcoded value - Stack Overflow

    Mar 16, 2023 · It's not possible, Databricks just scans entire output for occurences of secret values and replaces them with " [REDACTED]". It is helpless if you transform the value. For example, like you …

  3. Is there a way to use parameters in Databricks in SQL with parameter ...

    Sep 29, 2024 · EDIT: I got a message from Databricks' employee that currently (DBR 15.4 LTS) the parameter marker syntax is not supported in this scenario. It might work in the future versions. …

  4. Microsoft Fabric Unity Catalog mirroring from Azure Databricks fails

    Dec 28, 2025 · I’m trying to mirror an Azure Databricks Unity Catalog table into Microsoft Fabric using the Mirrored Azure Databricks catalog feature. I’ve validated that: Unity Catalog permissions are …

  5. REST API to query Databricks table - Stack Overflow

    Jul 24, 2022 · Is databricks designed for such use cases or is a better approach to copy this table (gold layer) in an operational database such as azure sql db after the transformations are done in pyspark …

  6. Convert string to date in databricks SQL - Stack Overflow

    May 30, 2021 · Use Databricks Datetime Patterns. According to SparkSQL documentation on the Databricks website, you can use datetime patterns specific to Databricks to convert to and from date …

  7. Do you know how to install the 'ODBC Driver 17 for SQL Server' on a ...

    Apr 4, 2020 · By default, Azure Databricks does not have ODBC Driver installed. Run the following commands in a single cell to install MS SQL ODBC Driver on Azure Databricks cluster.

  8. How to import own modules from repo on Databricks?

    Dec 7, 2022 · I have connected a Github repository to my Databricks workspace, and am trying to import a module that's in this repo into a notebook also within the repo. The structure is as such: …

  9. Installing multiple libraries 'permanently' on Databricks' cluster ...

    Feb 28, 2024 · Easiest is to use databricks cli 's libraries command for an existing cluster (or create job command and specify appropriate params for your job cluster) Can use the REST API itself, same …

  10. Databricks: How do I get path of current notebook?

    Databricks is smart and all, but how do you identify the path of your current notebook? The guide on the website does not help. It suggests: %scala dbutils.notebook.getContext.notebookPath res1: ...