LogoLogo
HomeAcademyLoginTry for free
  • Welcome
  • What's new
    • Q2 2025
    • Q1 2025
    • Q4 2024
    • Q3 2024
    • Q2 2024
    • Q1 2024
    • Q4 2023
    • Q3 2023
    • Q2 2023
    • Q1 2023
    • Q4 2022
    • Q3 2022
  • FAQs
    • Accounts
    • Migration to the new platform
    • User & organization setup
    • General
    • Builder
    • Workflows
    • Data Observatory
    • Analytics Toolbox
    • Development Tools
    • Deployment Options
    • CARTO Basemaps
    • CARTO for Education
    • Support Packages
    • Security and Compliance
  • Getting started
    • What is CARTO?
    • Quickstart guides
      • Connecting to your data
      • Creating your first map
      • Creating your first workflow
      • Developing your first application
    • CARTO Academy
  • CARTO User Manual
    • Overview
      • Creating your CARTO organization
      • CARTO Cloud Regions
      • CARTO Workspace overview
    • Maps
      • Data sources
        • Simple features
        • Spatial Indexes
        • Pre-generated tilesets
        • Rasters
        • Defining source spatial data
        • Managing data freshness
        • Changing data source location
      • Layers
        • Point
          • Grid point aggregation
          • H3 point aggregation
          • Heatmap point aggregation
          • Cluster point aggregation
        • Polygon
        • Line
        • Grid
        • H3
        • Raster
        • Zoom to layer
      • Widgets
        • Formula widget
        • Category widget
        • Pie widget
        • Histogram widget
        • Range widget
        • Time Series widget
        • Table widget
      • SQL Parameters
        • Date parameter
        • Text parameter
        • Numeric parameter
        • Publishing SQL parameters
      • Interactions
      • Legend
      • Basemaps
        • Basemap selector
      • AI Agents
      • SQL analyses
      • Map view modes
      • Map description
      • Feature selection tool
      • Search locations
      • Measure distances
      • Exporting data
      • Download PDF reports
      • Managing maps
      • Sharing and collaboration
        • Editor collaboration
        • Map preview for editors
        • Map settings for viewers
        • Comments
        • Embedding maps
        • URL parameters
      • Performance considerations
    • Workflows
      • Workflow canvas
      • Results panel
      • Components
        • Aggregation
        • Custom
        • Data Enrichment
        • Data Preparation
        • Generative AI
        • Input / Output
        • Joins
        • Parsers
        • Raster Operations
        • Spatial Accessors
        • Spatial Analysis
        • Spatial Constructors
        • Spatial Indexes
        • Spatial Operations
        • Statistics
        • Tileset Creation
        • BigQuery ML
        • Snowflake ML
        • Google Earth Engine
        • Google Environment APIs
        • Telco Signal Propagation Models
      • Data Sources
      • Scheduling workflows
      • Sharing workflows
      • Using variables in workflows
      • Executing workflows via API
      • Temporary data in Workflows
      • Extension Packages
      • Managing workflows
      • Workflows best practices
    • Data Explorer
      • Creating a map from your data
      • Importing data
        • Importing rasters
      • Geocoding data
      • Optimizing your data
    • Data Observatory
      • Terminology
      • Browsing the Spatial Data Catalog
      • Subscribing to public and premium datasets
      • Accessing free data samples
      • Managing your subscriptions
      • Accessing your subscriptions from your data warehouse
        • Access data in BigQuery
        • Access data in Snowflake
        • Access data in Databricks
        • Access data in Redshift
        • Access data in PostgreSQL
    • Connections
      • Google BigQuery
      • Snowflake
      • Databricks
      • Amazon Redshift
      • PostgreSQL
      • CARTO Data Warehouse
      • Sharing connections
      • Deleting a connection
      • Required permissions
      • IP whitelisting
      • Customer data responsibilities
    • Applications
    • Settings
      • Understanding your organization quotas
      • Activity Data
        • Activity Data Reference
        • Activity Data Examples
        • Activity Data Changelog
      • Users and Groups
        • Inviting users to your organization
        • Managing user roles
        • Deleting users
        • SSO
        • Groups
        • Mapping groups to user roles
      • CARTO Support Access
      • Customizations
        • Customizing appearance and branding
        • Configuring custom color palettes
        • Configuring your organization basemaps
        • Enabling AI Agents
      • Advanced Settings
        • Managing applications
        • Configuring S3 Bucket for Redshift Imports
        • Configuring OAuth connections to Snowflake
        • Configuring OAuth U2M connections to Databricks
        • Configuring S3 Bucket integration for RDS for PostgreSQL Exports in Builder
        • Configuring Workload Identity Federation for BigQuery
      • Data Observatory
      • Deleting your organization
    • Developers
      • Managing Credentials
        • API Base URL
        • API Access Tokens
        • SPA OAuth Clients
        • M2M OAuth Clients
      • Named Sources
  • Data and Analysis
    • Analytics Toolbox Overview
    • Analytics Toolbox for BigQuery
      • Getting access
        • Projects maintained by CARTO in different BigQuery regions
        • Manual installation in your own project
        • Installation in a Google Cloud VPC
        • Core module
      • Key concepts
        • Tilesets
        • Spatial indexes
      • SQL Reference
        • accessors
        • clustering
        • constructors
        • cpg
        • data
        • http_request
        • import
        • geohash
        • h3
        • lds
        • measurements
        • placekey
        • processing
        • quadbin
        • random
        • raster
        • retail
        • routing
        • s2
        • statistics
        • telco
        • tiler
        • transformations
      • Guides
        • Running queries from Builder
        • Working with Raster data
      • Release notes
      • About Analytics Toolbox regions
    • Analytics Toolbox for Snowflake
      • Getting access
        • Native App from Snowflake's Marketplace
        • Manual installation
      • Key concepts
        • Spatial indexes
        • Tilesets
      • SQL Reference
        • accessors
        • clustering
        • constructors
        • data
        • http_request
        • import
        • h3
        • lds
        • measurements
        • placekey
        • processing
        • quadbin
        • random
        • raster
        • retail
        • s2
        • statistics
        • tiler
        • transformations
      • Guides
        • Running queries from Builder
        • Working with Raster data
      • Release Notes
    • Analytics Toolbox for Databricks
      • Getting access
        • Personal (former Single User) cluster
        • Standard (former Shared) cluster
      • Reference
        • lds
        • tiler
      • Guides
      • Release Notes
    • Analytics Toolbox for Redshift
      • Getting access
        • Manual installation in your database
        • Installation in an Amazon Web Services VPC
        • Core version
      • Key concepts
        • Tilesets
        • Spatial indexes
      • SQL Reference
        • clustering
        • constructors
        • data
        • http_request
        • import
        • lds
        • placekey
        • processing
        • quadbin
        • random
        • s2
        • statistics
        • tiler
        • transformations
      • Guides
        • Running queries from Builder
      • Release Notes
    • Analytics Toolbox for PostgreSQL
      • Getting access
        • Manual installation
        • Core version
      • Key concepts
        • Tilesets
        • Spatial Indexes
      • SQL Reference
        • h3
        • quadbin
        • tiler
      • Guides
        • Creating spatial index tilesets
        • Running queries from Builder
      • Release Notes
    • CARTO + Python
      • Installation
      • Authentication Methods
      • Visualizing Data
      • Working with Data
        • How to work with your data in the CARTO Data Warehouse
        • How to access your Data Observatory subscriptions
        • How to access CARTO's Analytics Toolbox for BigQuery and create visualizations via Python notebooks
        • How to access CARTO’s Analytics Toolbox for Snowflake and create visualizations via Python notebooks
        • How to visualize data from Databricks
      • Reference
    • CARTO QGIS Plugin
  • CARTO for Developers
    • Overview
    • Key concepts
      • Architecture
      • Libraries and APIs
      • Authentication methods
        • API Access Tokens
        • OAuth Access Tokens
        • OAuth Clients
      • Connections
      • Data sources
      • Visualization with deck.gl
        • Basemaps
          • CARTO Basemap
          • Google Maps
            • Examples
              • Gallery
              • Getting Started
              • Basic Examples
                • Hello World
                • BigQuery Tileset Layer
                • Data Observatory Tileset Layer
              • Advanced Examples
                • Arc Layer
                • Extrusion
                • Trips Layer
            • What's New
          • Amazon Location
            • Examples
              • Hello World
              • CartoLayer
            • What's New
        • Rapid Map Prototyping
      • Charts and widgets
      • Filtering and interactivity
      • Summary
    • Quickstart
      • Make your first API call
      • Visualize your first dataset
      • Create your first widget
    • Guides
      • Build a public application
      • Build a private application
      • Build a private application using SSO
      • Visualize massive datasets
      • Integrate CARTO in your existing application
      • Use Boundaries in your application
      • Avoid exposing SQL queries with Named Sources
      • Managing cache in your CARTO applications
    • Reference
      • Deck (@deck.gl reference)
      • Data Sources
        • vectorTableSource
        • vectorQuerySource
        • vectorTilesetSource
        • h3TableSource
        • h3QuerySource
        • h3TilesetSource
        • quadbinTableSource
        • quadbinQuerySource
        • quadbinTilesetSource
        • rasterSource
        • boundaryTableSource
        • boundaryQuerySource
      • Layers (@deck.gl/carto)
      • Widgets
        • Data Sources
        • Server-side vs. client-side
        • Models
          • getFormula
          • getCategories
          • getHistogram
          • getRange
          • getScatter
          • getTimeSeries
          • getTable
      • Filters
        • Column filters
        • Spatial filters
      • CARTO APIs Reference
    • Release Notes
    • Examples
    • CARTO for React
      • Guides
        • Getting Started
        • Views
        • Data Sources
        • Layers
        • Widgets
        • Authentication and Authorization
        • Basemaps
        • Look and Feel
        • Query Parameters
        • Code Generator
        • Sample Applications
        • Deployment
        • Upgrade Guide
      • Examples
      • Library Reference
        • Introduction
        • API
        • Auth
        • Basemaps
        • Core
        • Redux
        • UI
        • Widgets
      • Release Notes
  • CARTO Self-Hosted
    • Overview
    • Key concepts
      • Architecture
      • Deployment requirements
    • Quickstarts
      • Single VM deployment (Kots)
      • Orchestrated container deployment (Kots)
      • Advanced Orchestrated container deployment (Helm)
    • Guides
      • Guides (Kots)
        • Configure your own buckets
        • Configure an external in-memory cache
        • Enable Google Basemaps
        • Enable the CARTO Data Warehouse
        • Configure an external proxy
        • Enable BigQuery OAuth connections
        • Configure Single Sign-On (SSO)
        • Use Workload Identity in GCP
        • High availability configuration for CARTO Self-hosted
        • Configure your custom service account
      • Guides (Helm)
        • Configure your own buckets (Helm)
        • Configure an external in-memory cache (Helm)
        • Enable Google Basemaps (Helm)
        • Enable the CARTO Data Warehouse (Helm)
        • Configure an external proxy (Helm)
        • Enable BigQuery OAuth connections (Helm)
        • Configure Single Sign-On (SSO) (Helm)
        • Use Workload Identity in GCP (Helm)
        • Use EKS Pod Identity in AWS (Helm)
        • Enable Redshift imports (Helm)
        • Migrating CARTO Self-hosted installation to an external database (Helm)
        • Advanced customizations (Helm)
        • Configure your custom service account (Helm)
    • Maintenance
      • Maintenance (Kots)
        • Updates
        • Backups
        • Uninstall
        • Rotating keys
        • Monitoring
        • Change the Admin Console password
      • Maintenance (Helm)
        • Monitoring (Helm)
        • Rotating keys (Helm)
        • Uninstall (Helm)
        • Backups (Helm)
        • Updates (Helm)
    • Support
      • Get debug information for Support (Kots)
      • Get debug information for Support (Helm)
    • CARTO Self-hosted Legacy
      • Key concepts
        • Architecture
        • Deployment requirements
      • Quickstarts
        • Single VM deployment (docker-compose)
      • Guides
        • Configure your own buckets
        • Configure an external in-memory cache
        • Enable Google Basemaps
        • Enable the CARTO Data Warehouse
        • Configure an external proxy
        • Enable BigQuery OAuth connections
        • Configure Single Sign-On (SSO)
        • Enable Redshift imports
        • Configure your custom service account
        • Advanced customizations
        • Migrating CARTO Self-Hosted installation to an external database
      • Maintenance
        • Updates
        • Backups
        • Uninstall
        • Rotating keys
        • Monitoring
      • Support
    • Release Notes
  • CARTO Native App for Snowflake Containers
    • Deploying CARTO using Snowflake Container Services
  • Get Help
    • Legal & Compliance
    • Previous libraries and components
    • Migrating your content to the new CARTO platform
Powered by GitBook
On this page
  • Case When
  • Cast
  • Columns to Array
  • Create Column
  • Drop Columns
  • Edit Schema
  • Extract from JSON
  • Find and Replace
  • Generate UUID
  • Geography to Geometry
  • Geometry to Geography
  • Hex Color Generator
  • Is not Null
  • Limit
  • Multi-col formula
  • Multi-row formula
  • Normalize
  • Order by
  • Poly Build
  • Poly Split
  • Remove Duplicated
  • Rename Column
  • Row Number
  • Sample
  • Select
  • Select Distinct
  • Simple Filter
  • Spatial Filter
  • ST SetSRID
  • Text to columns
  • Transpose / Unpivot
  • Unique
  • Where

Was this helpful?

Export as PDF
  1. CARTO User Manual
  2. Workflows
  3. Components

Data Preparation

Components to prepare your data for downstream analysis, this can include altering a table structure, re-ordering data, subsampling data, etc.

Case When

Description

This component generates column values that depend on a set of specified conditions.

Inputs

  • Source table [Table]

  • Conditional expressions: The UI of this component helps creating a conditional expression involving multiple columns and SQL operators. Each expression will produce a different result, as set on the component.

  • Result Column [Column]: Select a column that will contain the specified resulting value.

Cast

Description

This component casts the content of a column to a given type

Inputs

  • Source table [Table]

  • Column [Column]

  • New type [Selection]

Outputs

  • Result table [Table]

Columns to Array

Description

This component adds a new column with an array containing the values in a set of selected columns.

Inputs

  • Source table [Table]

  • Columns

  • Array column name

Outputs

  • Result table [Table]

Create Column

Description

This component creates a new table with an additional column computed using an expression.

Inputs

  • Source table [Table]

  • Name for new column [String]

  • Expression [String]

Outputs

  • Result table [Table]

External links

Drop Columns

Description

This component generates a new table with the same content as the input one, except one of its columns.

The component will fail if the column to remove is the only one in the input table.

Inputs

  • Source table [Table]

  • Column [Column]

Outputs

  • Result table [Table]

External links

Edit Schema

Description

This component simplifies the process of modifying table schemas. It allows to select specific columns, with the option to adjust their names and data types as required.

Inputs

  • Source table [Table]

  • Columns: The component's UI allows selecting a column, giving a new name and selecting a data type to cast the column.

Outputs

  • Result table [Table]

Extract from JSON

Description

This component creates a new column with values extracted from the JSON strings in another column. It uses the Data Warehouse syntax to specify the path to the key that needs to be extracted. See the documentation links below for more information.

Inputs

  • Source table [Table]

  • JSON column [Column]

  • JSON path [expression]

  • New column [Column]

Output

  • Result table

Find and Replace

Description

This component finds a string in one column of a table and replaces it with the specified value from another table.

As an alternative, columns from the lookup table can be added to the original table in those rows where the searched string is found. This is regulated by the Mode parameter.

Inputs

  • Source table [Table]

  • Find within column [Column]

  • Lookup table [Table]

  • Find value column [Column]

  • Replacement column [Column]

  • Find mode [Selection]

  • Case insensitive [Boolean]

  • Match whole word [Boolean]

  • Mode [Selection]

  • Columns to append [Column] [Multiple]: only used if Append field(s) to record mode is selected

Outputs

  • Result table [Table]

Generate UUID

Description

This component creates a new table with an additional UUID column named id.

Inputs

  • Source table [Table]

Outputs

  • Result table [Table]

Geography to Geometry

Description

This component converts a column from geography to geometry data type.

Inputs

  • Source table [Table]

  • geography column [Column]

Outputs

  • Result table [Table]

Geometry to Geography

Description

This component converts a column from geometry to geography data type.

Inputs

  • Source table [Table]

  • Geometry column [Column]

Outputs

  • Result table [Table]

Hex Color Generator

Description

This component create hex color for each distinct value of an input string column. Column values as NULL will be associated to grey value. The component generates a copy of the source with a new string column called: [name_of_input_col] + '_hex_color'.

Inputs

  • Source table [Table]

  • Column with category values

Outputs

  • Result table [Table]

Is not Null

Description

This component filters an input table using the presence or absence of null values in a given column.

Inputs

  • Source table [Table]

  • Column [Column]

Outputs

  • Not null values table [Table]

  • Null values table [Table]

Limit

Description

This component creates a new table with only the N first rows of the input table.

Inputs

  • Source table [Table]

  • Number of rows [Number]

Outputs

  • Result table [Table]

External links

Multi-col formula

Description

This component computes new values based on a given expression and a set of fields to apply the expression to. Use $a to refer to the value of the current column.

Inputs

  • Source table[Table]

  • Expression [String]. The expression to apply

  • Mode [Selection]. The mode used to put new values in the table

  • Prefix [String]. Only for the case of mode='Create new columns'

  • Columns [Column][Multiple]. The columns to apply the formula to

Outputs

  • Result table [Table]

Multi-row formula

Description

This component creates a new table containing a new column computed using a multi-row formula based on one or several input columns.

To refer to a value in the previous row, use {colname - 1} and to refer to a value in the next row, use {colname + 1}.

Inputs

  • Table[Table]

  • New column name [String]

  • New column type [Selection]

  • Expression [String]

  • Value for missing row values [Selection]

  • Column to sort by [Column]

  • Column to group by [Column]

Outputs

  • Result table [Table]

Normalize

Description

This component normalizes the values of a given column.

It adds a new column named '[column_name]_norm'.

Normalization can be computed as 0-1 values or as z-scores

Inputs

  • Source table [Table]

  • Column to normalize [Column]

  • Use z-scores [Boolean].

    • Disabled (default): When disabled, the resulting normalized values will range between 0 an 1.

Outputs

  • Result table [Table]

Order by

Description

This component generates a new table containing the rows of an input one sorted according to the values in one of its columns, and an optional second column.

Columns to use cannot be of type geometry.

Inputs

  • Table to order [Table]

  • Column to order by [Column]

  • Use descending order [Boolean]

  • Optional secondary column to order by [Column]

  • Use descending order in secondary column [Boolean]

Outputs

  • Result table [Table]

External links

Poly Build

Description

This component takes a group of spatial point objects and draws a polygon or polyline in a specific sort order to represent that group of points.

This component can also be used for spatial layer development by translating a collection of GPS data into polygon or polyline objects, where a polygon is a simple bounded region, such as a state boundary, and a polyline contains multiple line segments with any number of points between its start and endpoints, such as a river or road.

Inputs

  • Build Method [Selection]

  • Source table [Table]

  • Source Field [Column]

  • Source Field [Column]

  • Sequence Field [Column]

Outputs

  • Result table [Table]

Poly Split

Description

This component splits polygon or polyline objects into their component point, line, or region objects.

This is a very specialized component used for spatial layer development. A typical use of this component is to disaggregate complex regions that may contain more than one polygon or to separate a polyline into its individual nodes.

Inputs

  • Source table [Table]

  • Spatial Field [Column]

  • Split To [Selection]

Outputs

  • Result table [Table]

Remove Duplicated

Description

This component takes an input table and generates a new one in which duplicates rows from the input table have been removed.

Inputs

  • Source table [Table]

Outputs

  • Result table [Table]

Rename Column

Description

This component generates a new table with the same content as the input one, renaming one or multiple of its columns.

Inputs

  • Source table [Table]

  • Column to rename [Column]

  • New column name [String]

Outputs

  • Result table [Table]

External links

Row Number

Description

This component creates a new table with an additional column containing row numbers.

Inputs

  • Source table [Table]

Outputs

  • Result table [Table]

Sample

Description

This component generates a new table with a random sample of N rows from an input table.

Inputs

  • Source table [Table]

  • Number of rows to sample [Number]

Outputs

  • Result table [Table]

Select

Description

This component runs an arbitrary SELECT statement.

It generates a new table from a SELECT statement applied on the input table.

Inputs

  • Source table [Table]

  • SELECT statement [String]

Outputs

  • Result table [Table]

External links

Select Distinct

Description

This component generates a new table with the unique values that appear in a given column of an input table.

Inputs

  • Source table [Table]

  • Column [Column]

Outputs

  • Result table [Table]

External links

Simple Filter

Description

This component filters an input table according to a filter expression based on a single column.

It generates a new table with only the rows of the input table that meet the filter criteria and another one with those that do not meet it.

Inputs

  • Source table [Table]

  • Column [Column]

  • Operator [Selection]

  • Value [String]

Outputs

  • Table with rows that pass the filter [Table]

  • Table with rows that do not pass the filter [Table]

Spatial Filter

Description

This component filters an input table using a spatial predicate and a filter table.

It generates a new table with only the rows of the input table that meet the filter criteria and another one with those that do not meet it.

Inputs

  • Source table [Table]

  • Filter table [Table]

  • Geo column in source table [Column]

  • Geo column in filter table [Column]

  • Spatial predicate [Selection]

Outputs

  • Table with rows that pass the filter [Table]

  • Table with rows that do not pass the filter [Table]

ST SetSRID

Description

This component sets the SRID of a geo column

Inputs

  • Source table [Table]

  • Geo column [Column]

  • SRID [String]

Outputs

  • Result table [Table]

Text to columns

Description

This component adds new columns based on splitting the text string in a text column.

Inputs

  • Table [Table]

  • Column to split [Column]

  • Delimiters [String]

  • Mode [Selection]. Whether to add new columns or new rows with splitted strings

  • Number of new columns: Only used if mode = 'Split to columns'

  • Prefix for new column names [String]: Only used if mode = 'Split to columns'

  • Extra characters: What to do with extra characters if there are more tokens after dividing the string according to the delimiters than the ones defined in the 'Number of new columns' parameter

Outputs

  • Result table [Table]

Transpose / Unpivot

Description

This component rotates table columns into rows.

Inputs

  • Table to unpivot [Table]

  • Key columns [Column][Multiple]:The columns to use for identifying rows

  • Data columns [Column][Multiple]: The columns to use for key-value pairs

Outputs

  • Result table [Table]

Unique

Description

This component separates unique rows and duplicated rows. Unique rows are defined using the values in one or several columns of the input table, as defined by the user.

Inputs

  • Tables [Table]

  • Columns to find unique values [Column][Multiple]

Outputs

  • Table with unique rows [Table]

  • Table with duplicated rows [Table]

Where

Description

This component filters an input table according to a filter expression.

It generates a new table with only the rows of the input table that meet the filter criteria and another one with those that do not meet it.

Inputs

  • Source table [Table]

  • Filter expression [String]

Outputs

  • Table with rows that pass the filter [Table]

  • Table with rows that do not pass the filter [Table]

External links

  • Result table [Table]

PreviousData EnrichmentNextGenerative AI

Last updated 12 months ago

Was this helpful?

Enabled: The normalized value will be calcuated as a z-score or standard score: the number of standard deviations that the value is above or below the mean of the whole column. See .

BigQuery reference
BigQuery reference
BigQuery reference
Snowflake reference
Redshift reference
PostgreSQL reference
BigQuery reference
Snowflake reference
BigQuery reference
Snowflake reference
reference
BigQuery reference
Snowflake reference
Redshift reference
BigQuery reference
BigQuery reference
Snowflake reference
PostgreSQL reference
BigQuery reference
Redshift reference
BigQuery reference
Redshift reference
BigQuery reference
Snowflake reference