DEPLOYMENT

Sovereign where it matters.On-prem everywhere.

Three deployment models. Same AI, same product. What changes is where your data lives and where the inference runs.

THE 3 MODES

Same product, three different perimeters.

DocZoom is the same application across every configuration. What changes is where the app runs, where the AI inference runs, and who owns the hardware. The page does not ask you to choose: a sales call helps you map the mode to your compliance perimeter.

01 · ON-PREM

DocZoom On-Prem

A machine dedicated to you, hosted by us in EU data centers. No shared cloud: your team works on a dedicated machine, fully managed by us, with custom domain and brand. Nothing to operate, no hardware to buy, time-to-value measured in hours.

  • HostingA machine dedicated to you · EU data centers
  • AIInference in EU data centers, EU residency, zero retention
  • ICPProfessional firms, SMEs, scale-ups

For teams that want to start now without managing infrastructure.

02 · DEDICATED ON-PREM

DocZoom Dedicated On-Prem

DocZoom installed inside the customer's private cloud · AWS, Azure or GCP, EU region. The app runs in your perimeter, under your contract with the hyperscaler. AI inference local on your cloud GPUs, or via API · your call. We handle deployment and runbooks; you hold the keys.

  • HostingCustomer cloud (AWS / Azure / GCP EU region)
  • AILocal on your cloud GPUs, or via API
  • ICPEnterprise with private cloud, regulated, finance

For enterprises that already run a private cloud and want the deployment inside their own perimeter.

03 · SOVEREIGN

DocZoom Sovereign

Everything on the customer's physical hardware, inside their network. AI inference local on on-site GPUs (NVIDIA DGX small/medium or HP DL380 + 2× RTX 6000 Ada). Air-gappable: no outbound calls, no egress. The strictest mode, for ultra-sensitive data.

  • HostingCustomer physical hardware, on-premise
  • AILocal on customer GPUs, air-gappable
  • ICPDefense, intelligence, ultra-sensitive data

For organisations that cannot have a single bit of their data leave the rack.

Not sure which mode fits? You are not alone · even CIOs mix up Dedicated On-Prem and Sovereign. A 30-minute scoping call clears it up.

HARDWARE

Certified enterprise hardware.

For Sovereign, where the customer wants AI on their own rack, we rely on datacenter-grade enterprise hardware. Two reference architectures, picked by workload and budget.

NVIDIA DGX

SMALL · MEDIUM SOVEREIGN

NVIDIA DGX

NVIDIA integrated system for sovereign AI inference. We use it in Sovereign when the workload is small-to-medium. Desktop or rack form factor depending on size.

  • Fully local inference, zero outbound calls
  • Compact form factor, 1U/2U rack depending on size
  • Turnkey setup: arrives configured
  • Available on the Sovereign track (on the customer rack)

ENTERPRISE SOVEREIGN

HP DL380 + 2× NVIDIA RTX 6000 Ada

HP DL380 enterprise server with two NVIDIA RTX 6000 Ada Generation cards. Our reference for enterprise Sovereign: 96 GB combined VRAM, sufficient for production models on fleets up to hundreds of concurrent users.

  • 2× NVIDIA RTX 6000 Ada (96 GB combined VRAM)
  • Hardware cost ~€40-50K, scales to enterprise fleets
  • HP 5-year financing at preferential rates available
  • Hardware maintenance and spare parts backed by HP

The HP DL380 + RTX 6000 Ada option is available with HP 5-year financing at preferential rates: it turns a significant capex into a predictable monthly fee, aligned with the IT budget cycle.

THREE COMPONENTS · ONE TECHNOLOGY

Studio, Word and Italian databases run on every mode.

The three DocZoom components share the same RAG engine and the same knowledge base. No incompatibility across deployment modes: when you move from On-Prem to Sovereign, the products stay the same, the databases stay connected and the workflows keep running.

01 · CORE

Studio

The flagship product: semantic search across your knowledge base, document agents, RAG with verifiable citations. The backbone the other two plug into.

02 · DRAFTING

Word

Microsoft Word add-in for drafting agreements, deeds and contractual documents. Lives inside your Word, talks to the Studio knowledge base, keeps everything on your domain.

03 · DATABASES

Italian databases

Cassazione, Codes, AdE/INPS practice, CCNL and Italian legal sources already integrated. Answers from Studio and drafts in Word always cite the source, also in Dedicated On-Prem and Sovereign modes.

One brand, three components, three deploys. No incompatibility between tiers.

FAQ

Six questions, six answers.

The questions our sales team gets in the first 48 hours of an enterprise conversation.

On-Prem is a machine dedicated to you but hosted by us in EU data centers: the infrastructure is ours, we manage it, and you work with your brand and domain · AI inference runs in EU data centers, GDPR-safe. Sovereign instead runs on the customer's physical hardware, inside their network · DGX Spark or HP DL380 + RTX 6000: the iron is yours, the AI runs on your on-site GPUs, no egress. On-Prem gives you isolation and zero ops overhead; Sovereign gives full technical sovereignty but requires an ops team on your side.

NVIDIA Partner

NVIDIA TECHNOLOGY PARTNER

Reference architectures and AI deployment expertise on NVIDIA infrastructure across all three modes: from On-Prem (DGX managed by us) to Dedicated On-Prem (customer cloud) to Sovereign (DGX or HP DL380 + RTX 6000 Ada on the customer rack).

Which mode is right for you?

Talk to sales