LogoLogo
HomeAcademyLoginTry for free
  • Welcome
  • What's new
    • Q2 2025
    • Q1 2025
    • Q4 2024
    • Q3 2024
    • Q2 2024
    • Q1 2024
    • Q4 2023
    • Q3 2023
    • Q2 2023
    • Q1 2023
    • Q4 2022
    • Q3 2022
  • FAQs
    • Accounts
    • Migration to the new platform
    • User & organization setup
    • General
    • Builder
    • Workflows
    • Data Observatory
    • Analytics Toolbox
    • Development Tools
    • Deployment Options
    • CARTO Basemaps
    • CARTO for Education
    • Support Packages
    • Security and Compliance
  • Getting started
    • What is CARTO?
    • Quickstart guides
      • Connecting to your data
      • Creating your first map
      • Creating your first workflow
      • Developing your first application
    • CARTO Academy
  • CARTO User Manual
    • Overview
      • Creating your CARTO organization
      • CARTO Cloud Regions
      • CARTO Workspace overview
    • Maps
      • Data sources
        • Simple features
        • Spatial Indexes
        • Pre-generated tilesets
        • Rasters
        • Defining source spatial data
        • Managing data freshness
        • Changing data source location
      • Layers
        • Point
          • Grid point aggregation
          • H3 point aggregation
          • Heatmap point aggregation
          • Cluster point aggregation
        • Polygon
        • Line
        • Grid
        • H3
        • Raster
        • Zoom to layer
      • Widgets
        • Formula widget
        • Category widget
        • Pie widget
        • Histogram widget
        • Range widget
        • Time Series widget
        • Table widget
      • SQL Parameters
        • Date parameter
        • Text parameter
        • Numeric parameter
        • Publishing SQL parameters
      • Interactions
      • Legend
      • Basemaps
        • Basemap selector
      • AI Agents
      • SQL analyses
      • Map view modes
      • Map description
      • Feature selection tool
      • Search locations
      • Measure distances
      • Exporting data
      • Download PDF reports
      • Managing maps
      • Sharing and collaboration
        • Editor collaboration
        • Map preview for editors
        • Map settings for viewers
        • Comments
        • Embedding maps
        • URL parameters
      • Performance considerations
    • Workflows
      • Workflow canvas
      • Results panel
      • Components
        • Aggregation
        • Custom
        • Data Enrichment
        • Data Preparation
        • Generative AI
        • Input / Output
        • Joins
        • Parsers
        • Raster Operations
        • Spatial Accessors
        • Spatial Analysis
        • Spatial Constructors
        • Spatial Indexes
        • Spatial Operations
        • Statistics
        • Tileset Creation
        • BigQuery ML
        • Snowflake ML
        • Google Earth Engine
        • Google Environment APIs
        • Telco Signal Propagation Models
      • Data Sources
      • Scheduling workflows
      • Sharing workflows
      • Using variables in workflows
      • Executing workflows via API
      • Temporary data in Workflows
      • Extension Packages
      • Managing workflows
      • Workflows best practices
    • Data Explorer
      • Creating a map from your data
      • Importing data
        • Importing rasters
      • Geocoding data
      • Optimizing your data
    • Data Observatory
      • Terminology
      • Browsing the Spatial Data Catalog
      • Subscribing to public and premium datasets
      • Accessing free data samples
      • Managing your subscriptions
      • Accessing your subscriptions from your data warehouse
        • Access data in BigQuery
        • Access data in Snowflake
        • Access data in Databricks
        • Access data in Redshift
        • Access data in PostgreSQL
    • Connections
      • Google BigQuery
      • Snowflake
      • Databricks
      • Amazon Redshift
      • PostgreSQL
      • CARTO Data Warehouse
      • Sharing connections
      • Deleting a connection
      • Required permissions
      • IP whitelisting
      • Customer data responsibilities
    • Applications
    • Settings
      • Understanding your organization quotas
      • Activity Data
        • Activity Data Reference
        • Activity Data Examples
        • Activity Data Changelog
      • Users and Groups
        • Inviting users to your organization
        • Managing user roles
        • Deleting users
        • SSO
        • Groups
        • Mapping groups to user roles
      • CARTO Support Access
      • Customizations
        • Customizing appearance and branding
        • Configuring custom color palettes
        • Configuring your organization basemaps
        • Enabling AI Agents
      • Advanced Settings
        • Managing applications
        • Configuring S3 Bucket for Redshift Imports
        • Configuring OAuth connections to Snowflake
        • Configuring OAuth U2M connections to Databricks
        • Configuring S3 Bucket integration for RDS for PostgreSQL Exports in Builder
        • Configuring Workload Identity Federation for BigQuery
      • Data Observatory
      • Deleting your organization
    • Developers
      • Managing Credentials
        • API Base URL
        • API Access Tokens
        • SPA OAuth Clients
        • M2M OAuth Clients
      • Named Sources
  • Data and Analysis
    • Analytics Toolbox Overview
    • Analytics Toolbox for BigQuery
      • Getting access
        • Projects maintained by CARTO in different BigQuery regions
        • Manual installation in your own project
        • Installation in a Google Cloud VPC
        • Core module
      • Key concepts
        • Tilesets
        • Spatial indexes
      • SQL Reference
        • accessors
        • clustering
        • constructors
        • cpg
        • data
        • http_request
        • import
        • geohash
        • h3
        • lds
        • measurements
        • placekey
        • processing
        • quadbin
        • random
        • raster
        • retail
        • routing
        • s2
        • statistics
        • telco
        • tiler
        • transformations
      • Guides
        • Running queries from Builder
        • Working with Raster data
      • Release notes
      • About Analytics Toolbox regions
    • Analytics Toolbox for Snowflake
      • Getting access
        • Native App from Snowflake's Marketplace
        • Manual installation
      • Key concepts
        • Spatial indexes
        • Tilesets
      • SQL Reference
        • accessors
        • clustering
        • constructors
        • data
        • http_request
        • import
        • h3
        • lds
        • measurements
        • placekey
        • processing
        • quadbin
        • random
        • raster
        • retail
        • s2
        • statistics
        • tiler
        • transformations
      • Guides
        • Running queries from Builder
        • Working with Raster data
      • Release Notes
    • Analytics Toolbox for Databricks
      • Getting access
        • Personal (former Single User) cluster
        • Standard (former Shared) cluster
      • Reference
        • lds
        • tiler
      • Guides
      • Release Notes
    • Analytics Toolbox for Redshift
      • Getting access
        • Manual installation in your database
        • Installation in an Amazon Web Services VPC
        • Core version
      • Key concepts
        • Tilesets
        • Spatial indexes
      • SQL Reference
        • clustering
        • constructors
        • data
        • http_request
        • import
        • lds
        • placekey
        • processing
        • quadbin
        • random
        • s2
        • statistics
        • tiler
        • transformations
      • Guides
        • Running queries from Builder
      • Release Notes
    • Analytics Toolbox for PostgreSQL
      • Getting access
        • Manual installation
        • Core version
      • Key concepts
        • Tilesets
        • Spatial Indexes
      • SQL Reference
        • h3
        • quadbin
        • tiler
      • Guides
        • Creating spatial index tilesets
        • Running queries from Builder
      • Release Notes
    • CARTO + Python
      • Installation
      • Authentication Methods
      • Visualizing Data
      • Working with Data
        • How to work with your data in the CARTO Data Warehouse
        • How to access your Data Observatory subscriptions
        • How to access CARTO's Analytics Toolbox for BigQuery and create visualizations via Python notebooks
        • How to access CARTO’s Analytics Toolbox for Snowflake and create visualizations via Python notebooks
        • How to visualize data from Databricks
      • Reference
    • CARTO QGIS Plugin
  • CARTO for Developers
    • Overview
    • Key concepts
      • Architecture
      • Libraries and APIs
      • Authentication methods
        • API Access Tokens
        • OAuth Access Tokens
        • OAuth Clients
      • Connections
      • Data sources
      • Visualization with deck.gl
        • Basemaps
          • CARTO Basemap
          • Google Maps
            • Examples
              • Gallery
              • Getting Started
              • Basic Examples
                • Hello World
                • BigQuery Tileset Layer
                • Data Observatory Tileset Layer
              • Advanced Examples
                • Arc Layer
                • Extrusion
                • Trips Layer
            • What's New
          • Amazon Location
            • Examples
              • Hello World
              • CartoLayer
            • What's New
        • Rapid Map Prototyping
      • Charts and widgets
      • Filtering and interactivity
      • Summary
    • Quickstart
      • Make your first API call
      • Visualize your first dataset
      • Create your first widget
    • Guides
      • Build a public application
      • Build a private application
      • Build a private application using SSO
      • Visualize massive datasets
      • Integrate CARTO in your existing application
      • Use Boundaries in your application
      • Avoid exposing SQL queries with Named Sources
      • Managing cache in your CARTO applications
    • Reference
      • Deck (@deck.gl reference)
      • Data Sources
        • vectorTableSource
        • vectorQuerySource
        • vectorTilesetSource
        • h3TableSource
        • h3QuerySource
        • h3TilesetSource
        • quadbinTableSource
        • quadbinQuerySource
        • quadbinTilesetSource
        • rasterSource
        • boundaryTableSource
        • boundaryQuerySource
      • Layers (@deck.gl/carto)
      • Widgets
        • Data Sources
        • Server-side vs. client-side
        • Models
          • getFormula
          • getCategories
          • getHistogram
          • getRange
          • getScatter
          • getTimeSeries
          • getTable
      • Filters
        • Column filters
        • Spatial filters
      • CARTO APIs Reference
    • Release Notes
    • Examples
    • CARTO for React
      • Guides
        • Getting Started
        • Views
        • Data Sources
        • Layers
        • Widgets
        • Authentication and Authorization
        • Basemaps
        • Look and Feel
        • Query Parameters
        • Code Generator
        • Sample Applications
        • Deployment
        • Upgrade Guide
      • Examples
      • Library Reference
        • Introduction
        • API
        • Auth
        • Basemaps
        • Core
        • Redux
        • UI
        • Widgets
      • Release Notes
  • CARTO Self-Hosted
    • Overview
    • Key concepts
      • Architecture
      • Deployment requirements
    • Quickstarts
      • Single VM deployment (Kots)
      • Orchestrated container deployment (Kots)
      • Advanced Orchestrated container deployment (Helm)
    • Guides
      • Guides (Kots)
        • Configure your own buckets
        • Configure an external in-memory cache
        • Enable Google Basemaps
        • Enable the CARTO Data Warehouse
        • Configure an external proxy
        • Enable BigQuery OAuth connections
        • Configure Single Sign-On (SSO)
        • Use Workload Identity in GCP
        • High availability configuration for CARTO Self-hosted
        • Configure your custom service account
      • Guides (Helm)
        • Configure your own buckets (Helm)
        • Configure an external in-memory cache (Helm)
        • Enable Google Basemaps (Helm)
        • Enable the CARTO Data Warehouse (Helm)
        • Configure an external proxy (Helm)
        • Enable BigQuery OAuth connections (Helm)
        • Configure Single Sign-On (SSO) (Helm)
        • Use Workload Identity in GCP (Helm)
        • Use EKS Pod Identity in AWS (Helm)
        • Enable Redshift imports (Helm)
        • Migrating CARTO Self-hosted installation to an external database (Helm)
        • Advanced customizations (Helm)
        • Configure your custom service account (Helm)
    • Maintenance
      • Maintenance (Kots)
        • Updates
        • Backups
        • Uninstall
        • Rotating keys
        • Monitoring
        • Change the Admin Console password
      • Maintenance (Helm)
        • Monitoring (Helm)
        • Rotating keys (Helm)
        • Uninstall (Helm)
        • Backups (Helm)
        • Updates (Helm)
    • Support
      • Get debug information for Support (Kots)
      • Get debug information for Support (Helm)
    • CARTO Self-hosted Legacy
      • Key concepts
        • Architecture
        • Deployment requirements
      • Quickstarts
        • Single VM deployment (docker-compose)
      • Guides
        • Configure your own buckets
        • Configure an external in-memory cache
        • Enable Google Basemaps
        • Enable the CARTO Data Warehouse
        • Configure an external proxy
        • Enable BigQuery OAuth connections
        • Configure Single Sign-On (SSO)
        • Enable Redshift imports
        • Configure your custom service account
        • Advanced customizations
        • Migrating CARTO Self-Hosted installation to an external database
      • Maintenance
        • Updates
        • Backups
        • Uninstall
        • Rotating keys
        • Monitoring
      • Support
    • Release Notes
  • CARTO Native App for Snowflake Containers
    • Deploying CARTO using Snowflake Container Services
  • Get Help
    • Legal & Compliance
    • Previous libraries and components
    • Migrating your content to the new CARTO platform
Powered by GitBook
On this page
  • Setup requirements
  • Connecting to Databricks using OAuth (U2M)
  • Connecting to Databricks using OAuth (M2M)
  • Connecting to Databricks using Personal Access Tokens (PATs)
  • Advanced options
  • Requiring viewer credentials on shared Databricks OAuth U2M connections
  • IP Whitelisting

Was this helpful?

Export as PDF
  1. CARTO User Manual
  2. Connections

Databricks

PreviousSnowflakeNextAmazon Redshift

Last updated 2 months ago

Was this helpful?

CARTO can connect to your Databricks Data Warehouse, allowing you to use your data for building Maps, Workflows and custom applications. There are three ways to set up a Databricks connection.

Recommended methods:

  • : Users authenticate into Databricks using their individual Databricks credentials, generating an access token for each user. This is the recommended setup, but it needs to be configured by an Admin first.

  • : Users authorize unattended access to Databricks resources with a service principal. This method is ideal when developing applications or using service accounts.

Other methods:

  • : Connect to Databricks using a Personal Access Token. This method provides a straightforward setup without requiring an OAuth configuration, but the methods above represent a more secure strategy for production environments.

Databricks strongly recommends using OAuth over Personal Access Tokens. OAuth tokens are automatically refreshed by default and do not require the direct management of the access token, improving your security against token hijacking and unwanted access.

CARTO is a fully cloud-native platform that runs queries on your behalf to power maps, workflows, etc. We never create or maintain any copies of your data.


Once connected to your Databricks account, CARTO will push SQL queries that will be executed through your Databricks SQL Warehouse or create and run jobs on your Databricks All-purspose compute cluster. Currently, the level of support varies for different geospatial data formats:

Simple features (as WKB binaries)
H3 indexes

Maps * - Prepared tables

✅

✅

Maps * - SQL Queries

❌

✅

Workflows

✅

✅

(*) Maps in Builder, Data Explorer, Worflows previews and custom apps created with our developer Tools

Setup requirements

These requirements apply regardless of the authentication method used:


Connecting to Databricks using OAuth (U2M)

CARTO can connect to Databricks with OAuth user-to-machine (U2M) for interactive access to Databricks resources.

Head to the Connections settings from the side menu, click on Databricks and select Setup connection with OAuth U2M, which will open the new connection form. You will be taken to your Databricks login page where you will have to authenticate with your own personal credentials.

Once authenticated, you will be redirected to the new connection form. These are the fields you need to provide:

  • Name: The name for the connection you're creating.

  • Catalog: The Unity Catalog to use with this connection. This should be the catalog your Service Principal has access to and contains the data you want to use in CARTO.

  • SQL Warehouse: The SQL Warehouse that will be used to execute SQL queries. This is used to list resources, get data for widgets and fetch map tiles.

  • All-purpose compute: The All-purpose compute cluster that will be used to create and run jobs. This is used to create and run Workflows.

Connecting to Databricks using OAuth (M2M)

CARTO can connect to Databricks with OAuth machine-to-machine (M2M), which provides unattended access to your resources.

As a prerequisite, a Service Principal must be created in Databricks, as well as an OAuth Secret for that Service Principal. For detailed steps on how to do this, please follow Databricks' official guide:

Once you have created the Service Principal and its OAuth Secret, head to the Connections settings from the side menu, click on Databricks and select Setup connection with OAuth M2M. This will open the new connection form:

These are the fields you need to provide:

  • Name: The name for the connection you're creating.

  • Service Principal Client ID: The Client ID of the Service Principal you wish to use.

  • Service Principal Secret: The OAuth secret of the Service Principal you wish to use.

  • Catalog: The Unity Catalog to use with this connection. This should be the catalog your Service Principal has access to and contains the data you want to use in CARTO.

  • SQL Warehouse: The SQL Warehouse that will be used to execute SQL queries. This is used to list resources, get data for widgets and fetch map tiles.

  • All-purpose compute: The All-purpose compute cluster that will be used to create and run jobs. This is used to create and run Workflows.

Connecting to Databricks using Personal Access Tokens (PATs)

CARTO can connect to Databricks using a Personal Access Token, which provides access to resources at the Databricks Workspace level. To create a new token, follow the steps on Databrick's official documentation:

Then head to the Connections settings from the side menu, click on Databricks and select Connect using personal access token. This will open the new connection form:

These are the fields you need to provide:

  • Name: The name for the connection you're creating.

  • Token: The Personal Access Token. The connection will inherit permission and access privileges of the user that generates the token.

  • Catalog: The Unity Catalog to use with this connection.

  • SQL Warehouse: The SQL Warehouse that will be used to execute SQL queries. This is used to list resources, get data for widgets and fetch map tiles.

  • All-purpose compute: The All-purpose compute cluster that will be used to create and run jobs. This is used to create and run Workflows.


Advanced options

Connections can be set up with these advanced options:

  • Max number of concurrent queries: The maximum number of simultaneous queries that CARTO will send to Databricks in that connection.

  • Max query timeout: This sets the maximum allowed duration of queries that CARTO runs in Databricks in that connection.

Requiring viewer credentials on shared Databricks OAuth U2M connections

Databricks OAuth U2M connections can be set up to require viewer credentials. This means that instead of using the credentials of the user that created the connection, each user will have to provide their own credentials to use it.

To require viewer credentials on your Databricks OAuth U2M connection, head to the Connections section and click on Permissions and Sharing from the Connection card and then set the Share mode to Organization. By default, Viewer Credentials will be checked:

After the setup, other users (regardless of their role) will see a prompt like this every time they want to use the connection. This happens when creating/viewing maps that use that connection or when previewing the connection's data in the Data Explorer.

If a map is public, users won't be asked for their credentials even if the connection requires viewer credentials.

IP Whitelisting

CARTO connections require in your Databricks workspace.

acceleration must be enabled in the All-purpose compute clusters that you use with CARTO.

15.4 LTS is recommended. Minimum DBR required is 14.2.

For visualizing (as WKB binaries), tables need to be . In order to do so, your Databricks workspace needs to be enabled with Spatial SQL functions, which are currently in Private Preview.

The Databricks team has made available to request access to the functions. Please get in touch with them through the form to gain access to all Spatial SQL functions.

As a prerequisite, an organization Admin needs to create a Databricks OAuth integration first. Once this is done, Databricks OAuth U2M connections will be available to all users within the organization. .

Make sure that you follow the when creating a new Databricks connection.

Host: Your Databricks instance name URL, for example dbc-xxxxx113-0000.cloud.databricks.com or carto-data-science.cloud.databricks.com. Learn more about this on the official .

Make sure that you follow the when creating a new Databricks connection.

Host: Your Databricks instance name URL, for example dbc-xxxxx113-0000.cloud.databricks.com or carto-data-science.cloud.databricks.com. Learn more about this on the official .

Make sure that you follow the when creating a new Databricks connection.

Restrict this connection to only use Named Sources: When enabled, this connection will only work within apps that use , and will NOT work in Data Explorer, Builder and Workflows. This prevents the usage of arbitrary SQL in applications for this connection.

If you're using the cloud version of CARTO (SaaS), CARTO will connect to Databricks using a set of static IPs for each region. for your specific region.

Unity Catalog
Photon
Databricks Runtime Release (DBR)
simple features
prepared for visualization
this form
Read more about setting up a Databricks OAuth (U2M) integration
Authorize unattended access to Databricks resources with a service principal using OAuth
Databricks documentation
Databricks personal access token authentication
Databricks documentation
Named Sources
Check this guide to find the IPs you need to allow
What it means to be fully cloud native.
OAuth (U2M)
OAuth (M2M)
Personal Access Token (PAT)
setup requirements
setup requirements
setup requirements