← Back to all products

Multi-Cloud Lakehouse Blueprint

$69

Architecture blueprint and Terraform templates for Databricks on Azure + AWS with unified governance, cost management, and Delta Sharing.

📁 17 files🏷 v1.0.0
PythonTerraformJSONMarkdownYAMLAWSAzureDatabricksDelta LakeRedis

📁 File Structure 17 files

multi-cloud-lakehouse-blueprint/ ├── README.md ├── cicd/ │ └── multi-cloud-pipeline.yml ├── docs/ │ ├── architecture_decision_records.md │ ├── compliance_matrix.md │ ├── identity_federation.md │ ├── migration_guide.md │ └── network_architecture.md ├── dr/ │ └── disaster_recovery_patterns.md ├── terraform/ │ ├── aws/ │ │ ├── main.tf │ │ ├── outputs.tf │ │ └── variables.tf │ ├── azure/ │ │ ├── main.tf │ │ ├── outputs.tf │ │ └── variables.tf │ └── shared/ │ └── unity-catalog-multicloud.tf └── tools/ └── cost_comparison_model.py

📖 Documentation Preview README excerpt

Multi-Cloud Lakehouse Blueprint

Product by [Datanest Digital](https://datanest.dev)

Overview

The Multi-Cloud Lakehouse Blueprint is a production-ready architecture package for

organizations building or extending a Databricks Lakehouse across both Microsoft Azure

and Amazon Web Services (AWS). It provides battle-tested Terraform modules, CI/CD

pipelines, governance configurations, and operational runbooks that eliminate months

of cross-cloud integration work.

Who Is This For?

  • Platform engineers tasked with deploying Databricks workspaces across Azure and AWS
  • Data architects designing multi-cloud data mesh or lakehouse topologies
  • Cloud architects evaluating or implementing cross-cloud disaster recovery
  • Engineering managers who need cost models and architecture decision records to

justify multi-cloud investments to leadership

What's Included

Documentation

| Document | Description |

|----------|-------------|

| docs/architecture_decision_records.md | 12 ADRs covering when and why to adopt multi-cloud lakehouse |

| docs/network_architecture.md | Cross-cloud connectivity patterns — VPN, peering, Private Link |

| docs/identity_federation.md | Azure AD + AWS IAM federation with SCIM provisioning |

| docs/compliance_matrix.md | GDPR, CCPA, data residency requirements across jurisdictions |

| docs/migration_guide.md | Step-by-step guide to extend Azure-only deployments to multi-cloud |

Terraform Modules

| Module | Description |

|--------|-------------|

| terraform/azure/ | Databricks workspace on Azure with VNet injection and Private Link |

| terraform/aws/ | Databricks workspace on AWS with VPC, NAT, and PrivateLink |

| terraform/shared/unity-catalog-multicloud.tf | Unity Catalog with cross-cloud Delta Sharing |

CI/CD

| Pipeline | Description |

|----------|-------------|

| cicd/multi-cloud-pipeline.yml | Unified Azure DevOps pipeline deploying to both clouds |

Tools

| Tool | Description |

|------|-------------|

| tools/cost_comparison_model.py | Interactive Azure vs AWS DBU pricing comparison calculator |

Disaster Recovery

| Runbook | Description |

|---------|-------------|

| dr/disaster_recovery_patterns.md | Active-passive cross-cloud DR patterns and failover procedures |

Architecture Overview

... continues with setup instructions, usage examples, and more.

📄 Code Sample .py preview

tools/cost_comparison_model.py #!/usr/bin/env python3 """ Multi-Cloud Lakehouse Blueprint — Cost Comparison Model Datanest Digital — https://datanest.dev Interactive tool for comparing Databricks costs between Azure and AWS. Covers DBU pricing, compute, storage, networking, and total cost of ownership. """ from __future__ import annotations import argparse import json import sys from dataclasses import dataclass, field from enum import Enum from typing import Optional class WorkloadType(Enum): """Databricks workload types with associated DBU multipliers.""" JOBS_COMPUTE = "Jobs Compute" ALL_PURPOSE = "All Purpose" JOBS_LIGHT = "Jobs Light" SQL_CLASSIC = "SQL Classic" SQL_PRO = "SQL Pro" SQL_SERVERLESS = "SQL Serverless" DELTA_LIVE_TABLES_CORE = "DLT Core" DELTA_LIVE_TABLES_PRO = "DLT Pro" DELTA_LIVE_TABLES_ADVANCED = "DLT Advanced" MODEL_SERVING = "Model Serving" @dataclass class DBUPricing: """DBU pricing per workload type for a specific cloud.""" cloud: str rates: dict[WorkloadType, float] = field(default_factory=dict) # --------------------------------------------------------------------------- # Default pricing (list prices as of early 2026 — update as needed) # Prices are in USD per DBU # --------------------------------------------------------------------------- AZURE_DEFAULT_RATES: dict[WorkloadType, float] = { WorkloadType.JOBS_COMPUTE: 0.15, WorkloadType.ALL_PURPOSE: 0.55, # ... 434 more lines ...