Databricks
Sumble connects directly to your Databricks workspace to read your account data and write enriched results back as Delta tables.
How it works
Sumble uses a direct SQL connection to your Databricks workspace (via the Databricks SQL API). The integration is bi-directional:
Pull — Sumble queries your account list from a table you specify (e.g., main.customer_schema.accounts). Column mappings translate your field names to Sumble's schema. This tells Sumble which organizations to match and enrich.
Push — After enrichment, Sumble writes the results as Delta tables in your Databricks workspace:
Parquet files are generated from the enrichment pipeline
Files are uploaded to intermediate cloud storage
Delta tables are created in a timestamped schema in your workspace
Tables are atomically swapped into your output schema, so you always see a complete, consistent dataset
What gets delivered
Each pipeline run writes these tables to your output schema:
Enriched organizations — your accounts matched to Sumble organizations with all configured enrichment columns
Alerts feed — intent signals generated for your accounts
Reference data — technologies, job functions, countries, and other lookup tables
Tables are replaced atomically with each run. Previous versions are retained briefly for rollback if needed.
Update cadence
Data is refreshed daily after the Sumble data pipeline completes.
Setup
Databricks integration requires an enterprise plan. To get started:
Provide Sumble with your Databricks workspace hostname, SQL warehouse HTTP path, and a Personal Access Token
Specify the input table containing your account list and the output schema for enriched data
Sumble configures column mappings for your account fields (account ID, name, URL, etc.)
Choose which enrichments to include
Last updated