HockeyStack
HomeLive DemoBook a DemoLogin
  • Getting Started
    • 👋Welcome to the Docs!
    • Product Onboarding
    • HockeyStack Implementation Scope: Reporting Product
  • Release Notes
    • May 5, 2025
    • April 21, 2025
    • April 14, 2025
    • April 7, 2025
    • March 31, 2025
    • March 24, 2025
    • March 14, 2025
    • March 6, 2025
    • February 28, 2025
    • February 17, 2025
  • Guides
    • ❓FAQ
      • Merging in HockeyStack
      • Why has my data changed?
      • Viewing form submissions by page
      • A touchpoint's influence on conversion rates
      • Average number of touchpoints
      • Self-reported attribution
      • Trend of Engagement Score
      • How do I see which individuals / companies are included in a metric?
      • What touchpoints get credit in attribution?
      • What object/integration is this field pulling from?
      • What is "Source"?
      • What is "UTM Source"?
      • What is "UTM Medium"?
      • What is an Action?
      • How can I add Salesforce Task object into Defined Properties?
      • How can I track offline events in HubSpot?
      • How can I use LinkedIn Impressions and Engagements in my reports?
      • Measuring sales and marketing penetration in an account list
      • Measuring number of engaged contacts per company
      • Offline conversions for ad platforms
      • Tracking progression on targets
      • Building a Campaign / Asset Grouping property
      • Percentage of high quality job titles by Channel
      • Measuring conversion rates
      • Building a goal that shows open opportunities
      • Number report: Funnel stages influenced by different types of marketing touchpoints
      • Best Practices for Lifecycle Tracking in Salesforce / HubSpot
      • Adding HubSpot form fills to defined properties
      • Does HockeyStack website pixel track US States?
      • Measuring Time Between Two Goals in HockeyStack
      • What is the HubSpot "email bounces" action?
      • Hiding Fields from your CRM in HockeyStack
      • How long does it take for a relation mapping to be ready to use?
      • Can I integrate multiple LinkedIn ads accounts?
      • Can I use Zapier for integrations with HockeyStack?
      • I created a new field in Salesforce (SFDC), but I don’t see it in HockeyStack. What should I do?
      • Using two similar fields in one breakdown
      • Why can't I map back to property?
      • GA4 vs. HockeyStack Website Data Tracking
      • How does HockeyStack deduplicate accounts?
      • How do Table Totals Work: Campaign vs Campaign Group?
      • Can I create one field that calculates the total ad spend + SFDC campaign spend?
      • HubSpot: Can I filter a goal on X object by Y object fields?
      • LinkedIn Impressions: Different Ways of Measuring
      • How to define Engaged Accounts and Engaged People?
      • Building a Campaign Grouping property
    • 🖥️Dashboard Building Guides
      • Business Overview Dashboard
      • CMO Dashboard
      • Website Analytics Dashboard
      • Paid Ads Dashboard
      • Google Ads Dashboard
      • LinkedIn Ads Dashboard
      • In-Person Events
      • ABM Live-Demo
      • Content/Organic Dashboard
      • Dashboards from Labs Reports
        • LinkedIn Ads Benchmarks
        • Google Ads Benchmarks
        • Q1 2024 Recap
        • G2 Impact 2024 Report
        • Website Benchmarks
  • Documentation
    • The HockeyStack Data Model
    • 🎯Goals
      • Funnel Stages Goals
      • Form Fill Goals
      • Page View Goals
      • Click Goals
      • Finding Out a Button's CSS Selector
      • Goals on the Task Object
      • Building an All Touchpoints (Channel) Goal
    • Defined Properties
    • Track Date Properties
    • 📊Reports
      • Building a Basic Report
      • Journeys Use Cases
      • Customer Touchpoint Hierarchy
      • Sequences
      • Lift Reports
      • Lift Analysis vs. Multi-Touch Attribution
      • Types of report filters and when to use them
      • Attribution Models
      • Attribution Lookback
      • Defining Custom Attribution Weights
      • Importing a Google Sheet to use as a Goal Column
      • Advanced Attribution Models
    • 🖥️Dashboards
    • Dashboard Filters
      • When to use AND vs. OR logic?
      • Using Regex
    • 🌠Journeys
      • Syncing journeys to CRM and Slack
    • 🥇Golden Paths
    • Funnels
    • Attribution Funnel
    • 👥Segments
    • ⚙️Settings
      • Account Reset Guide
      • Auto-tagging of URLs
      • Data Categorization in HockeyStack
      • Team Sharing
      • Tracking Multiple Domains
      • Excluding Users
      • Reporting Configuration
      • Multi-Factor Authentication
    • Advanced Data Connections
      • Account List Import
      • Property Relation Mappings
      • Sync Spend
      • Syncing spend from offline channels and campaigns
    • 🔃Audience Syncs
    • Send View updates to Webhooks
    • Odin AI
      • HockeyStack AI: Security, Privacy, and Responsible Use
  • DataSyncs
    • Connecting your Warehouse
      • Authenticate Snowflake
      • Authenticate Google Sheets
      • Authenticate BigQuery
      • Authenticate S3
        • Use an S3 User
        • Use an IAM Role
    • Configure a DataSync Import
    • Configure a DataSync Export
      • Data Export Schema
        • Raw Actions Export Schema
  • Integrations
    • Website Tracker
      • Google Tag Manager
      • WordPress
      • React
      • Troubleshooting
      • Reverting to Cookie-Based Tracking
      • Identifying Users
      • Tracking Custom Goals
    • Ad Platforms
      • LinkedIn Ads
      • Bing Ads
      • Capterra Ads
      • Google Ads
      • Facebook Ads
      • Tiktok Ads
      • Twitter Ads
      • StackAdapt Ads
      • Reddit Ads
      • AdRoll Ads
    • Analytics & Data Warehouse
      • Snowflake
      • Amazon Redshift
      • Google Bigquery
      • Amazon S3
      • Azure Databricks
    • CRMs
      • Salesforce
        • Properties Pulled from Salesforce
        • Salesforce Pulled Objects List
        • Sending Data to Salesforce
      • HubSpot
        • HubSpot Pulled Objects List
    • SSO
      • Azure AD
      • Google Workspace
      • Okta
    • ABM
      • Qualified
      • 6sense
      • Demandbase
      • Clearbit
      • Rollworks
      • G2 Intent
      • Stackadapt
    • Marketing Automation
      • Marketo
        • How to Find Your Marketo Account Details
        • Marketo Pulled Objects List
      • Pardot
      • HubSpot
        • HubSpot Pulled Objects List
    • Other Integrations
      • Calendly
      • Drift
      • Okta
      • Segment
      • Customer.io
  • Setting up your Data for import
    • Import Custom Actions
    • Import Website Actions
    • Import Properties
    • Import Metadata
  • Technical Details
    • ↖️Website Tracking
      • How Website Tracking Works
      • Cookieless Tracking
      • Bot Traffic
      • Privacy Policy
      • GDPR Compliance
    • ⚙️Data Processing from Integrations
    • 🧮Data Cleaning
  • Account Intelligence
    • ☕Getting Started
      • HockeyStack Implementation Scope: Account Intelligence Product
      • Salesforce
        • Salesforce Permissions
        • Salesforce iFrame Installation
        • Salesforce Sync Fields
    • 🏗️Workflows
      • Creating a Workflow
      • List of Workflows
      • Starter Workflow
      • Recurring Workflow Runs
      • Nodes
        • Transformations
          • Condition
          • AI for Accounts
          • Contact Discovery
          • Contact Enrichment
          • Branching
        • Destinations
          • Salesforce
          • HubSpot
          • Outreach
          • StackAdapt
          • Salesloft
          • LinkedIn
          • Pardot
    • 👀Views
      • Create a New View
    • 🔢Scoring
      • Data
Powered by GitBook
On this page
  • Overview
  • Prerequisites & Credentials
  • Ingesting Data from Azure Databricks into HockeyStack
  • Data Requirements for Ingestion
  • Exporting Data from HockeyStack to Azure Databricks
  • How Export Works
  • Sync Options
  • Default Recommendations
  1. Integrations
  2. Analytics & Data Warehouse

Azure Databricks

PreviousAmazon S3NextCRMs

Last updated 4 months ago

Overview

Azure Databricks is a data lakehouse platform that many revenue operations and data teams use to store, process, and analyze valuable customer data. HockeyStack can integrate directly with Azure Databricks to both ingest data from your Databricks environment and export processed data back into it.

Prerequisites & Credentials

Before setting up the integration, gather the following credentials and share them securely with our support team. Be sure to review the for any additional details that may apply:

  • DATABRICKS_TOKEN: Your personal access token (PAT)

  • DATABRICKS_HOST: The Databricks host URL associated with your workspace

Please provide these details via a secure channel so we can establish a trusted connection with your environment.

Ingesting Data from Azure Databricks into HockeyStack

HockeyStack transforms your Databricks data into “actions” that can be viewed and analyzed within our platform’s activity timeline. Each action typically includes:

  • Action Name: A descriptor for what happened (e.g. “Monthly Revenue”).

  • Action Date(s): The date(s) associated with the action (e.g. 2024-01-01).

  • Entity Identifier: An identifier (e.g., email, company domain, CRM record ID) that links the action to a known person or company in HockeyStack.

  • Additional Attributes: Details like revenue amount, currency, or any other metrics that provide deeper insights.

Example Action Record:

action_name = Monthly Revenue
revenue_date = 2024-01-01
revenue_amount = 100
revenue_currency = USD
account_id = 183114939199

Data Requirements for Ingestion

To ensure a smooth ingestion process, the table(s) in Azure Databricks you plan to sync must meet the following criteria:

  • Incremental Sync Timestamp: At least one datetime column representing the “last modified date” of each record. HockeyStack uses this column to determine which records are new or updated since the last sync.

  • Action Name Column: At least one column should serve as the “action name” to categorize or identify the type of event or record.

  • Action Date Column(s): One or more columns should represent the date(s) on which the action occurred.

  • Entity Identifier Column: At least one column must contain an identifier (e.g., email, CRM ID) that matches identifiers known to HockeyStack. This allows actions to be tied back to the correct person or company in your system.

Once you’ve identified or prepared a table that meets these requirements, share the schema details with our support team to begin the ingestion setup.

Exporting Data from HockeyStack to Azure Databricks

HockeyStack can also push data into your Azure Databricks instance. This allows you to incorporate HockeyStack’s enriched analytics data back into your broader data ecosystem for advanced analytics, modeling, or reporting.

How Export Works

  • Scheduled Syncs: By default, HockeyStack runs daily export jobs that can be customized in frequency.

  • Data Formatting: Our workers query HockeyStack’s database, format the data as needed, and push it into the specified destination table(s) in Azure Databricks.

  • Schema Planning: Before initiating exports, our support team will propose a table structure and work with you to confirm the data columns, formats, and fields needed.

Sync Options

  • Incremental Sync: Push only new or updated records since the last export.

  • Recurring Full Sync: Clear the designated Databricks table and repopulate it entirely each run.

Default Recommendations

  • Daily Cronjob: Clears and resyncs the table once per day.

  • Data Format: We typically mirror HockeyStack’s raw data format — a single denormalized table containing a timeline of activity for each person and company. This format can vary per customer due to the range of data fields we ingest from different sources.


If you have any questions or need guidance on setup, data mapping, or optimizing your integration, our support team is here to help. By connecting Azure Databricks with HockeyStack, you’ll unlock powerful insights and streamlined workflows that help drive better decision-making and more efficient data operations.

Databricks documentation