Model Context Protocol (MCP) finally gives AI models a way to access the business data needed to make them really useful at work. CData MCP Servers have the depth and performance to make sure AI has access to all of the answers.
Try them now for free →Bridge Databricks Connectivity with Apache NiFi
Access and process Databricks data in Apache NiFi using the CData JDBC Driver.
Apache NiFi supports powerful and scalable directed graphs of data routing, transformation, and system mediation logic. When paired with the CData JDBC Driver for Databricks, NiFi can work with live Databricks data. This article describes how to connect to and query Databricks data from an Apache NiFi Flow.
With built-in optimized data processing, the CData JDBC driver offers unmatched performance for interacting with live Databricks data. When you issue complex SQL queries to Databricks, the driver pushes supported SQL operations, like filters and aggregations, directly to Databricks and utilizes the embedded SQL engine to process unsupported operations client-side (often SQL functions and JOIN operations). Its built-in dynamic metadata querying allows you to work with and analyze Databricks data using native data types.
About Databricks Data Integration
Accessing and integrating live data from Databricks has never been easier with CData. Customers rely on CData connectivity to:
- Access all versions of Databricks from Runtime Versions 9.1 - 13.X to both the Pro and Classic Databricks SQL versions.
- Leave Databricks in their preferred environment thanks to compatibility with any hosting solution.
- Secure authenticate in a variety of ways, including personal access token, Azure Service Principal, and Azure AD.
- Upload data to Databricks using Databricks File System, Azure Blog Storage, and AWS S3 Storage.
While many customers are using CData's solutions to migrate data from different systems into their Databricks data lakehouse, several customers use our live connectivity solutions to federate connectivity between their databases and Databricks. These customers are using SQL Server Linked Servers or Polybase to get live access to Databricks from within their existing RDBMs.
Read more about common Databricks use-cases and how CData's solutions help solve data problems in our blog: What is Databricks Used For? 6 Use Cases.
Getting Started
Connecting to Databricks Data in Apache NiFi
- Download the CData JDBC Driver for Databricks installer, unzip the package, run the .exe file to install the driver.
Copy the CData JDBC Driver JAR file (and license file if it exists), cdata.jdbc.databricks.jar (and cdata.jdbc.databricks.lic), to the Apache NiFi lib subfolder, for example, C:\nifi-1.3.0-bin\nifi-1.3.0\lib.
On Windows, the default location for the CData JDBC Driver is C:\Program Files\CData\CData JDBC Driver for Databricks.
Start Apache NiFi by running the run-nifi.bat file in bin subfolder, for example, C:\nifi-1.3.0-bin\nifi-1.3.0\bin.
(OR)
Use the command prompt to navigate to the particular directory and run the run-nifi.bat file for example:
cd C:\nifi-1.3.0-bin\nifi-1.3.0\bin .\run-nifi.bat
Navigate to the Apache NiFi UI in your web browser: It should be https://localhost:8443/nifi.
Note: If users are utilizing an older version of Apache NiFi, they should access it via http://localhost:8080/nifi. In earlier versions, HTTP was the protocol employed. However, in the most recent version, HTTPS is the standard. By default, HTTP operates on port 8080, while HTTPS uses port 8443.
When accessing Apache NiFi via a URL, it prompts you to enter a username and password for login.
To retrieve login credentials, users should consult the 'App.log' file located within the log directory of their NiFi installation. This file typically contains the necessary details for accessing the NiFi interface.
- Click the (Configuration) button from the Operate sidebar under the "NiFi Flow" process group.
- In the NiFi Flow Configuration page, navigate to the "Controller Services" tab.
- Click the button to create a new controller service.
- In the Add Controller Service page, select "DBCPConnectionPool", and then click Add.
- In the Configure Controller Service page, navigate to the "Properties" tab.
- Click the gear button to configure the new DBCPConnection Pool.
Fill in the properties:
Property Value Database Connection URL jdbc:databricks:Server=127.0.0.1;Port=443;TransportMode=HTTP;HTTPPath=MyHTTPPath;UseSSL=True;User=MyUser;Password=MyPassword; Database Driver Class Name cdata.jdbc.databricks.DatabricksDriver Database Driver Location(s) Path to your Apache NiFi's lib folder where the JAR files are present. Built-in Connection String Designer
For assistance in constructing the JDBC URL, use the connection string designer built into the Databricks JDBC Driver. Either double-click the JAR file or execute the JAR file from the command-line.
java -jar cdata.jdbc.databricks.jar
Fill in the connection properties and copy the connection string to the clipboard.
To connect to a Databricks cluster, set the properties as described below.
Note: The needed values can be found in your Databricks instance by navigating to Clusters, and selecting the desired cluster, and selecting the JDBC/ODBC tab under Advanced Options.
- Server: Set to the Server Hostname of your Databricks cluster.
- HTTPPath: Set to the HTTP Path of your Databricks cluster.
- Token: Set to your personal access token (this value can be obtained by navigating to the User Settings page of your Databricks instance and selecting the Access Tokens tab).
- To establish a connection and execute a select query, simply drag and drop the processor (indicated by the yellow highlight) into the workspace.
- Select the 'ExecuteSQL' processor and click the 'Add' button to make it visible in the workspace.
- Click on the added processor (ExecuteSQL)to open the connection page.
- Click on the Verify Properties button and verify to test the connection.
- Ensure that all the verification results are successful, and then click on APPLY.
- Activate the connection by clicking the small button at the right corner of the screen.
Your Databricks data is now available for use in Apache NiFi. For example, you can use the DBCPConnection Pool as the source for a QueryDatabaseTable processor (shown below).

Download a free, 30-day trial of the CData JDBC Driver for Databricks and start working with your live Databricks data in Apache NiFi. Reach out to our Support Team if you have any questions.