Last updated: Apr 13, 2026
HEALTH TECH VENDOR
IMPLEMENTATION
To populate your Microsoft Azure repository with healthcare data from an EHR system via Redox (and then to feed that data into Databricks for analytics), you must configure a specific Redox cloud destination. A Redox destination represents where a message is delivered (e.g., like the address in the “To” line of an email header). Learn more about connecting Redox to your cloud repository.
You’ll need to perform some steps in your cloud product(s) and some in Redox. You can perform Redox setup in our dashboard or with the Redox Platform API.
- Establish a connection with your preferred EHR system. Learn how to request a connection.
- Complete your Azure (with Data Lake) and Databricks configuration before creating your Redox destination. Save any downloads with secret values, since you’ll need to enter some of these details into the Redox dashboard.
- Grant access to Redox from Azure (and any other cloud product) to authorize Redox to push data to your cloud repository.
- Navigate to the Microsoft Azure dashboard and log in. Review Azure’s quickstart guide to get started.
- Create an application through Microsoft Entra ID. Review Azure’s help article. This is where you’ll get a client ID and tenant ID, which you’ll need for Redox setup later.
- Create a new secret for your application. This is where you’ll get client secret value, which you’ll need for Redox setup later.
- Create a new storage account. Set the primary service to the Data Lake Storage option.
- Add a new container. You’ll need the name of the container for Redox setup later.
- Assign the appropriate Blob Data role (i.e., must at least have write permissions) to the application you created in step #2.
Next, create a cloud destination in your Redox organization. When the EHR system sends healthcare data to Redox, we push it on to your configured Azure + Databricks cloud destination.
- For the select destination step, follow the instructions for creating a cloud destination.
- From the Product type field, select Databricks.
- For the configure destination step, populate these fields. Then click the Next button.
- Storage account name: Enter the name of the storage account you created in Azure. Locate this value in the Azure dashboard.
- Container name: Enter the name of the container you created in Azure. Locate this value in the Azure container configuration.
- File name prefix (optional): Enter any prefix you want prepended to new files when they’re created in the Data Lake container. Add / to put the files in a subdirectory. For example, redox/ puts all the files in the redox directory.
- For the auth credential step, either a drop-down list of existing auth credentials displays or a new auth credential form opens. Learn how to create an auth credential for OAuth 2.0 Two-legged.Token endpoint URLExisting or new auth credential
- For the verify step, follow the instructions for verifying a destination.
- In your terminal, prepare the /v1/authcredentials request.
- Specify these values in the request.
- Locate the clientId and clientSecret value in the Microsoft Azure dashboard.Example: Create auth credential for Azure + Databricksjson
- You should get a successful 200 response and a payload populated with the details of the new auth credential.
- In your terminal, prepare the /v1/environments/{environmentId}/destinations request.
- Specify these values in the request.
- Set authCredential to the auth credential ID from the response you received in step #4.
- Populate cloudProviderSettings with the settings below (adjust values based on the storage account and container setup in Azure configuration).
- Enter databricks as the productId.
- The fileNamePrefix is optional. If specified, the filename format will be the prefix you define appended by the file path in the Data Lake container. You can append / after the prefix name to indicate a directory path.Example: Values for Azure + Databricks cloudProviderSettingsjson
- You should get a successful 200 response with a payload populated with the details of the new Microsoft Azure + Databricks cloud destination. Specifically, the verified status of the destination should be set to true.
- Your new destination will now be able to receive messages. Redox pushes data to the Data Lake storage account as a JSON file, which is ingested into Microsoft Azure.