With this connector, Hasura allows you to instantly create a real-time GraphQL API on top of your data models in Databricks. This connector supports Databricks's functionalities listed in the table below, allowing for efficient and scalable data operations. Additionally, users benefit from all the powerful features of Hasura’s Data Delivery Network (DDN) platform, including query pushdown capabilities that delegate query operations to the database, thereby enhancing query optimization and performance.
This connector implements the Data Connector Spec.
Below, you'll find a matrix of all supported features for the Databricks connector:
Feature | Supported | Notes |
---|---|---|
Native Queries + Logical Models | ❌ | |
Native Mutations | ❌ | |
Simple Object Query | ✅ | |
Filter / Search | ✅ | |
Simple Aggregation | ✅ | |
Sort | ✅ | |
Paginate | ✅ | |
Table Relationships | ❌ | |
Views | ✅ | |
Remote Relationships | ❌ | |
Custom Fields | ❌ | |
Mutations | ❌ | |
Distinct | ❌ | |
Enums | ❌ | |
Naming Conventions | ❌ | |
Default Values | ❌ | |
User-defined Functions | ❌ |
- Create a Hasura Cloud account
- Please ensure you have the DDN CLI and Docker installed
- Create a supergraph
- Create a subgraph
With the context set for an existing subgraph, initialize the connector:
ddn connector init -i
When the wizard runs, you'll be prompted to enter the following env vars necessary for your connector to function:
Name | Description | Required |
---|---|---|
DATABRICKS_JDBC_URL | JDBC URL of your Databricks DB | Yes |
DATABRICKS_SCHEMA | Databricks Schema | Yes |
DATABRICKS_CATALOG | Databricks Catalog | Yes |
Note: While entering the JDBC URL, ensure that the JDBC URL is in accordance with the DataBricks JDBC URL format, as specified here.
After the CLI initializes the connector, you'll need to:
- Introspect the source.
- Add your models, commands, and relationships.
- Create a new build.
- Test it by running your project along with the connector.
The Hasura Cassandra connector is available under the Apache License 2.0.