Skip to main content

Enterprise (VPC Deployment)

For organizations that require data to never leave their network, HatiData Enterprise deploys the data plane directly inside your AWS, GCP, or Azure VPC. The control plane communicates exclusively over PrivateLink -- no public internet exposure.

Architecture

Your VPC                                  HatiData VPC
+-----------------------------------+ +---------------------------+
| | | |
| AI Agents / Applications | | Control Plane (Axum) |
| | | | - Auth (JWT/API key) |
| v | | - Billing |
| HatiData Proxy (:5439) | | - Policy management |
| - Postgres wire protocol | | - Audit |
| - SQL transpilation | | - Dashboard API |
| - DuckDB execution engine | | |
| - L1 RAM / L2 NVMe cache | +---------------------------+
| | | ^
| v | |
| Object Storage (your bucket) | PrivateLink (no public internet)
| - Iceberg-format tables | |
| - Encrypted at rest (KMS) | ---------+
| |
+-----------------------------------+

Key properties:

  • Data plane in your VPC -- The proxy, DuckDB engine, and caches run on your infrastructure
  • No data egress -- Query data never crosses VPC boundaries
  • PrivateLink connectivity -- Control plane communication uses VPC endpoints, not the public internet
  • Your storage, your keys -- Data stored in your S3/GCS/Azure Blob buckets, encrypted with your KMS keys

Cloud Provider Support

HatiData Enterprise supports all three major cloud providers:

ProviderData PlaneStorageEncryptionTerraform Modules
AWSEC2 Auto Scaling (i3 NVMe)S3 (Iceberg)AWS KMSterraform/aws/ (8 modules)
GCPGKE Standard or Cloud RunGCS (Iceberg)Cloud KMSterraform/gcp/ (11 modules)
AzureAKS or Container InstancesAzure Blob (Iceberg)Azure Key Vaultterraform/azure/ (6 modules)

Infrastructure (AWS Example)

HatiData Enterprise deploys via Terraform modules:

ModulePurpose
vpcVPC, subnets, security groups, NAT gateway
computeEC2 Auto Scaling Group (i3 instances with NVMe)
storageS3 buckets (data lake, audit), KMS encryption keys
iamIAM roles, instance profiles, cross-account access
control_planePrivateLink endpoint, VPC endpoint service
monitoringCloudWatch, Prometheus, alerting
marketplaceAWS Marketplace metering integration
shadow_modeQuery replay and compatibility testing

Compute Tiers

EnvironmentInstance TypeNVMe StoragePricing
Developmenti3.large475 GBSpot instances
Stagingi3.xlarge950 GBSpot instances
Productioni3.2xlarge1.9 TBOn-demand

NVMe storage is used for the L2 disk cache, providing sub-millisecond reads for frequently accessed data. On instance termination, NVMe volumes are cryptographically wiped via cryptsetup luksErase.

Shadow Mode

Before migrating production workloads, Shadow Mode lets you validate compatibility without risking your existing pipelines:

  1. Upload your existing query history
  2. Replay queries against HatiData in parallel with your current warehouse
  3. Compare results, latency, and costs
  4. Review a compatibility report showing which queries pass, which need adjustment
# Start shadow mode
curl -X POST -H "Authorization: Bearer $JWT" \
https://api.hatidata.com/v1/organizations/$ORG_ID/shadow-mode

# Upload query history
curl -X POST -H "Authorization: Bearer $JWT" \
-F "file=@query_history.csv" \
https://api.hatidata.com/v1/organizations/$ORG_ID/shadow-mode/upload

# Start replay
curl -X POST -H "Authorization: Bearer $JWT" \
https://api.hatidata.com/v1/organizations/$ORG_ID/shadow-mode/replay

# Get comparison report
curl -H "Authorization: Bearer $JWT" \
https://api.hatidata.com/v1/organizations/$ORG_ID/shadow-mode/report

Shadow Mode is particularly useful when migrating from Legacy Cloud Warehouses. The compatibility report identifies queries that need adjustment and provides suggested rewrites.

IP Protection (Shield)

Enterprise deployments include HatiData Shield -- a four-layer defense-in-depth system:

  1. Black Box Binary -- Symbol stripping, LTO, string obfuscation, anti-debug detection
  2. Heartbeat License -- JWT-based tokens with agent-aware feature gating
  3. Locked Instance -- No SSH/SSM access, encrypted EBS, AMI-only delivery
  4. Environment Binding -- Cloud metadata validation of account, project, or subscription and region

Security Features

Enterprise deployments include all security features described in the Security Overview:

  • RBAC with 6 roles (ServiceAccount, Developer, Analyst, Auditor, Admin, Owner)
  • ABAC with 10-attribute evaluation context and compiled policy sets
  • Row-level security with agent-aware placeholders
  • Column masking with role-based redaction levels
  • Hash-chained audit logs (SHA-256) with tamper detection
  • JIT access grants with time-bounded privilege escalation
  • Federated authentication (AWS STS, Azure AD, Google Cloud Identity)
  • Multi-tenancy with automatic tenant isolation and cross-tenant JOIN prevention

SLAs and Support

TierSLASupport
Standard99.9% uptimeBusiness hours email, 24h response
Premium99.95% uptime24/7 Slack, 4h response, named TAM
Mission Critical99.99% uptime24/7 phone, 1h response, dedicated team

Getting Started

Enterprise deployment is customized to your infrastructure. To get started:

  1. Contact sales -- Email sales@hatidata.com or visit hatidata.com
  2. Architecture review -- We map your VPC topology, data sources, and compliance requirements
  3. Shadow Mode trial -- Validate compatibility with your existing queries
  4. Terraform deployment -- We provide the modules; you run terraform apply in your account
  5. Go live -- Cut over agents and applications to your in-VPC HatiData proxy

Terraform Quick Start

Once you have access to the Terraform modules, deployment follows a standard workflow:

# Clone the modules
git clone https://github.com/HatiOS-AI/HatiData-Enterprise.git
cd HatiData-Enterprise/terraform/aws # or /gcp or /azure

# Configure
cp terraform.tfvars.example terraform.tfvars
# Edit terraform.tfvars with your account details, region, VPC ID, etc.

# Deploy
terraform init
terraform plan
terraform apply

The Terraform modules create all required infrastructure: VPC endpoints, compute instances, storage buckets, IAM roles, monitoring, and the PrivateLink connection to the control plane.

Migrating from Cloud

Upgrading from Cloud to Enterprise preserves your data, schemas, policies, and API keys. The migration process:

  1. Terraform provisions the data plane in your VPC
  2. Data is replicated from HatiData-managed storage to your S3/GCS/Azure Blob bucket
  3. DNS is updated to point to your in-VPC proxy
  4. The old cloud endpoint is decommissioned

Zero-downtime migration is supported with a brief period of dual-write during cutover.

Next Steps

Stay in the loop

Product updates, engineering deep-dives, and agent-native insights. No spam.