Best way to connect to a databricks datamart for further exploration in PowerBI
Situation: Databricks is used as enterprise data platform. In reality local teams are moving forward with a different speed than the global organization is. To support this but still keeping control on the "golden source" we're setting up an…
![](https://techprofile.blob.core.windows.net/images/wK5E1j8Ql0SGo0Ox0mr4rg.png?8D8D19)
![](https://techprofile.blob.core.windows.net/images/wK5E1j8Ql0SGo0Ox0mr4rg.png?8D8D19)
How do I orchestrate ML model retraining periodically?
I have to retrain every month or so a PyTorch Model trained on data obtained from processing tables sitting in Azure Data Lake Storage gen 1. So far, I have the following building blocks: A Databricks notebook that does the ETL job of…
What is the use of oldest-time-to-consider param in Jobs API?
Hi, I haven't found any documentation around what this value "oldest-time-to-consider": "1457570074236" is used for in the Databricks Job API. Can someone please direct me to the documentation that talks about the significance of…
Can you write multiple streaming queries(same schema, different input sources) into same Azure storage without overwriting?
Hi, I have to this requirement to write multiple streaming queries(same schema, different input sources) into same Azure blob delta lake gen 3 storage without overwriting. I need the data to co-exist in the same write directory, say like in 'append'…
Is there any way to do Custom dynamic mapping of different number of columns in dataflow or any other options to achieve this?
My source (CSV file in ADLS) has header record(3 columns) , detail records(5 columns) and trailer record(2 columns) . The header record has less number of columns than the detail records. When I try to convert this csv file to parquet, i m getting the…
MLOps using Azure Databricks & Azure ML - question on data prep for model inference and retraining.
I am using this blog (https://databricks.com/blog/2020/10/13/using-mlops-with-mlflow-and-azure.html) to set-up MLOps using Azure Databricks & Azure ML. As mentioned in the blog, we deploy MLflow model into an Azure ML environment using the built in…
How to use ARM template to restrict publicBlobAccess to managed Databricks storage account
In our organisation we are required to disable publicBlobAccess and enable TLS1_2 as minimum version on all storage accounts. Preferably we also use StorageV2 type instead of BlobStorage. When we create Databricks workspace, using ARM template, the…
How to query 3rd party Azure DataLake Gen2 and only store the results
First, what I am trying to do is I want to query and aggregate raw JSON files stored in a 3rd party's Azure Data Lake (Gen2) and store those aggregates in my own data lake or relation db. I do not want to physically copy all of those raw JSON files…
Linkedin connectivity with Azure
Hi, Is there a way to extract existing content from a Linkedin company page using Azure? Thanks
Azure repo - Can't create or push tag
I host a project on Azure DevOps Repositories and I would like to create several git tag for project version release. I'm a GitKraken user so I added a new tag and push it to origin but this error occurs: So I checked my permission on Azure…
![](https://techprofile.blob.core.windows.net/images/c510ec930c4d47c495df17dd914d1681.jpg)
The file updated in databricks is not reflecting in Azure Portal
I have a databrick workspace, where I read a file from Azure blob storage, updated a file and uploaded it in another Azure blob storage space. Now when I access that file through any databricks workspace, I can see the file and access the content of it.…
![](https://techprofile.blob.core.windows.net/images/VfQFAmOikEWfBHko2XlWTA.png?8D7F33)
I HAVE ERROR WHEN AZURE DATABRICKS WRITE COSMOS DB
I am trying to write a spark dataframe from azure databricks to a cosmos db database and I have this error Py4JJavaError: An error occurred while calling o753.save. : org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage…
Deletion of managed MLflow Artifacts
When using Workspace experiments in Azure Databricks with the default managed MLflow artifact location dbfs:/databricks/mlflow-tracking configured, we see the following message when deleting an MLflow experiment run: Deleted runs are restorable for…
Unable to delete folder in databricks "DBFS://
Hi, When I run the command %fs ls '/' in the results I see a folder path as "dbfs://" and name as "/". and tried to run the command in the notebook %fs ls '//' I get the java error and even not able to delete the folder. Please…
using streaming batch for multiple operations
I am new to spark and DataBricks and was trying to look for a solution where I can utilize a batch from a eventhub stream to accomplish multiple business logic but could not find any guidance. Stream I get from EventHub is a CDC stream from multiple…
![](https://techprofile.blob.core.windows.net/images/VfQFAmOikEWfBHko2XlWTA.png?8D7F33)
Databricks readstream writestream to Azure Synapse
I am having an issue on writing stream to Azure synapse with the following error . let's have a look and see if there is idea ?
![](https://techprofile.blob.core.windows.net/images/VfQFAmOikEWfBHko2XlWTA.png?8D7F33)
How to set security permissions to Databases in databricks through Notebooks
We are stuck on the way to set security permissions to Databases by using Notebooks %sql. At first, let me explain our situations and settings. We run the following code on Notebooks: %sql CREATE DATABASE X ; GRANT USAGE ON DATABASE X TO…
Databricks Pyspark exception handling best practices
Hi, In the current development of pyspark notebooks on Databricks, I typically use the python specific exception blocks to handle different situations that may arise. I am wondering if there are any best practices/recommendations or patterns to handle…
Azure Databricks Cluster
Hi, Created new cluster in Databricks (QA environment) Afte that when i try to click on data tab inroder to create database getting below error checked that cluster is up and running.
![](https://techprofile.blob.core.windows.net/images/VfQFAmOikEWfBHko2XlWTA.png?8D7F33)
Quota limit hit on tutorial notebook
I'm attempting to launch a default cluster (min 2, max 8) on the premium trial account in order to run 01-The-Databricks-Environment. I haven't been able to run any operations in the notebook. Receiving this error: Azure error code:…
![](https://techprofile.blob.core.windows.net/images/VfQFAmOikEWfBHko2XlWTA.png?8D7F33)