Skip to main content

Databricks Reference

Reference material for Databricks Publishing -- data type mappings, CLI arguments, and querying.

Data Type Mapping

Source types are mapped to Databricks types automatically.

PostgreSQL to Databricks

PostgreSQL TypeDatabricks Type
INTEGER, INT4INT
BIGINT, INT8BIGINT
SMALLINT, INT2SMALLINT
NUMERIC(p,s)DECIMAL(p,s)
REAL, FLOAT4FLOAT
DOUBLE PRECISIONDOUBLE
VARCHAR(n), TEXTSTRING
DATEDATE
TIMESTAMPTIMESTAMP_NTZ
TIMESTAMPTZTIMESTAMP
BOOLEANBOOLEAN
BYTEABINARY

SQL Server to Databricks

SQL Server TypeDatabricks Type
INTINT
BIGINTBIGINT
SMALLINTSMALLINT
TINYINTTINYINT
DECIMAL(p,s)DECIMAL(p,s)
FLOATDOUBLE
REALFLOAT
VARCHAR(n), NVARCHAR(n)STRING
DATEDATE
DATETIME, DATETIME2TIMESTAMP_NTZ
DATETIMEOFFSETTIMESTAMP
BITBOOLEAN
VARBINARYBINARY

CLI Reference

Databricks Publishing Arguments

OptionTypeDescription
--publish_target IDStringCredential ID for Databricks publishing (required)
--publish_schema_pattern PATTERNStringSchema naming pattern (default: {schema})
--publish_table_pattern PATTERNStringTable naming pattern (default: {table})
--databricks_table_type TYPEEnumexternal or managed (default: external)
--databricks_catalog CATALOGStringOverride catalog from credentials
--n_jobs NIntegerParallel workers for table creation (default: 1)

Querying Tables

Databricks SQL:

SELECT * FROM lakexpress_catalog.tpch_1.customer LIMIT 10;

Databricks Notebooks (Python):

df = spark.table("lakexpress_catalog.tpch_1.customer")
df.display()

Databricks Notebooks (SQL):

%sql
SELECT * FROM lakexpress_catalog.tpch_1.customer LIMIT 10

Delta Table Features (Managed Tables Only)

Time Travel:

SELECT * FROM lakexpress_catalog.tpch_1.customer VERSION AS OF 1;

Table History:

DESCRIBE HISTORY lakexpress_catalog.tpch_1.customer;

Optimize:

OPTIMIZE lakexpress_catalog.tpch_1.customer;

See Also

Copyright © 2026 Architecture & Performance.