TonoForge™ — The Engine of Distributed AI Factories | Tonomia
Tonomia AI Infrastructure
TonoForge™
The Engine of
Distributed AI Factories

TonoForge™ deploys 2× racks of enterprise-grade GPU compute — choose AMD Instinct MI355X (128 GPUs) or NVIDIA GB300 NVL72 (144 GPUs) — from MiTAC, Supermicro or Dell, inside a single 300 kW containerised AI factory.

300 kWPower
2 RacksPer Unit
144GPUs Max
2.2 EFFP4 Peak
5–20wDeploy
TonoForge™ — 300 kW AI Factory
Photo
TonoForge™ Container — Exterior
Video
TONOFORGE™ — DUAL-RACK AI FACTORY ARCHITECTURE ENERGY INSolar / Wind / Grid150–300 kW BATTERYLFP AmplificationPeak Shaving RACK A64× MI355Xor 72× B300 GPU RACK B64× MI355Xor 72× B300 GPU TonoFabric™RL Orchestration 800G NETWORKFiber Backbone END USERS100K+ Users HEAT RECOVERY → ADJACENT BUILDINGS PUE 1.05 MITAC MR1100LSUPERMICRO SRS / AS-4126DELL POWEREDGE XE9712 BATTERY AMPLIFICATION · LIQUID COOLING · HEAT RECOVERY · SOVEREIGN AI
300 kW
Power Envelope
2× Racks
Per Container
128–144
GPUs Total
2.2 EF
FP4 Peak
~42 TB
HBM3E Memory
GPU Configurations

Choose Your Compute — AMD or NVIDIA

Each TonoForge™ unit houses 2× full racks of GPU compute, delivered pre-integrated and factory-tested. Choose between AMD Instinct MI355X or NVIDIA GB300 NVL72 platforms, sourced from MiTAC, Supermicro, or Dell.

Both configurations deliver enterprise-grade AI training and inference capacity with liquid cooling, 800 Gb/s networking, and full TonoFabric™ orchestration.

  • AMD path: 2× 64-GPU MI355X racks — 128 GPUs, 36.8 TB HBM3E, 161 PFLOPS FP4
  • NVIDIA path: 2× GB300 NVL72 racks — 144 GPUs, 42 TB HBM3e, 2.2 EXAFLOPS FP4
  • Rack vendors: MiTAC MR1100L · Supermicro AS-4126GS / SRS-NVL72 · Dell XE9712
  • Centralised orchestration via TonoFabric™ with RL-based optimisation
Compare ConfigurationsIndustry Benchmark
GPU ACCELERATOR SPECIFICATIONS AMD INSTINCT MI355X CDNA 4 · 3nm MEMORY288 GB HBM3E BANDWIDTH8 TB/s FP4 PEAK (MXFP4)~10 PFLOPS / GPU FP8 PEAK~5 PFLOPS / GPU TDP1,400W (DLC required) ARCHITECTURE256 CUs · 8 XCDs · 256 MB Infinity Cache INTERCONNECTPCIe 5.0 x16 · 4th Gen Infinity Fabric DATATYPESFP64 FP32 FP16 BF16 FP8 MXFP6 MXFP4 PER 8-GPU PLATFORM: 2.3 TB HBM3E · 80.5 PFLOPS FP4 · 6 TB DDR5 · Broadcom P2200G NVIDIA B300 (BLACKWELL ULTRA) Blackwell · 4nm MEMORY288 GB HBM3e (12-hi) BANDWIDTH8 TB/s FP4 PEAK (DENSE)15 PFLOPS / GPU FP8 PEAK5 PFLOPS / GPU TDP1,400W (DLC required) NVLink 51.8 TB/s per GPU (900 GB/s uni) NETWORKINGConnectX-8 · 800 Gb/s per GPU CPUNVIDIA Grace ARM (72 cores per CPU) PER NVL72 RACK: 72 GPUs · 21 TB HBM3e · 1.1 EXAFLOPS FP4 · 36 Grace CPUs · 17 TB LPDDR5X BOTH GPUs: 288 GB HBM3e · 8 TB/s · 1,400W TDP · DLC COOLING
AMD MI355X RACK — 64 GPUs PER RACK 48U G4826Z5 #18× MI355X · 2× EPYC 9005 G4826Z5 #28× MI355X · 2× EPYC 9005 G4826Z5 #38× MI355X · 2× EPYC 9005 G4826Z5 #48× MI355X · 2× EPYC 9005 G4826Z5 #5–84× more servers (32 GPUs) 400GbE SWITCH · 64-PORT MGMT SWITCH · 32-PORT DIRECT LIQUID COOLING · CDU POWER SHELF · 33kW × 5 PER RACK TOTALS GPUs64× AMD MI355X MEMORY18.4 TB HBM3E FP4 PEAK80.5 PFLOPS (MXFP4) FORM48U EIA · 8× 4U Servers · DLC MiTAC MR1100L-64355X-02 Supermicro AS-4126GS-TNMR SERVER SPECS (G4826Z5) CPU2x EPYC 9755 (Turin) RAM24x DDR5-6400 (6 TB) GPU MEM2.3 TB HBM3E / server NICPensando Pollara 400 ADAPTERBroadcom P2200G BUSPCIe 5.0 UBB 2.0 COOLINGCold plate DLC FORM4U per server (8x=32U) SWITCH64p/32p 400GbE PSU6x 3.3kW Titanium
NVIDIA GB300 NVL72 — 72 GPUs PER RACK 48U COMPUTE TRAY 1–416 GPUs COMPUTE TRAY 5–816 GPUs COMPUTE TRAY 9–1216 GPUs COMPUTE TRAY 13–1824 GPUs NVLink SWITCH TRAYS1.8 TB/s / GPU ConnectX-8 SuperNIC · 800G 36× BlueField DPUs FULL LIQUID COOLING · CDU POWER SHELF · 8× 33kW 36× NVIDIA Grace ARM CPUs PER RACK TOTALS GPUs72× NVIDIA B300 Blackwell Ultra MEMORY21 TB HBM3e FP4 PEAK1.1 EXAFLOPS (dense FP4) FORM48U · 18 Compute Trays · DLC POWER132 kW (8× 33kW shelf) Supermicro SRS-GB300-NVL72 Dell PowerEdge XE9712 MiTAC / Ingrasys RACK SPECS (SRS-GB300-NVL72) NODES18x 1U ARS-121GL-NB3 GPU MEM21 TB HBM3e SYS MEM17 TB LPDDR5X NVLink9x switch, 1.8 TB/s/GPU NETWORKCX-8 800Gb/s + BF-3 DPU POWER8x 33kW shelf (132kW) CDU250kW in-rack, N+1 pumps STORAGE144x E1.S PCIe 5.0 DIMS600x1068x2236mm
TONOFORGE™ — CIRCULAR ENERGY ARCHITECTURE · PUE ~1.05 PUE~1.05 SOLARRenewable PPA WINDCo-located GRID150 kW backup LFP BATTERY2× Amplification AI COMPUTE128–144 GPUsMI355X or GB300 HEAT OUTPUT→ Adjacent buildingsRevenue stream NVMe STORAGEHigh-speed I/O BATTERY AMPLIFICATION · HEAT RECOVERY · RENEWABLE CO-LOCATION · NEAR-PERFECT EFFICIENCY NO GRID UPGRADE REQUIRED · PEAK SHAVING · ZERO-WASTE THERMAL LOOP
Feature 01
Battery Amplification Strategy

TonoForge™ deploys battery storage amplification to solve the grid bottleneck — one of the biggest blockers for AI infrastructure. By pairing compute with on-site LFP energy storage, TonoForge™ delivers peak AI performance without requiring expensive grid upgrades.

Heat waste is recovered and redistributed to adjacent buildings, creating a circular energy system with a PUE close to 1.05 — near-theoretical perfection.

  • Deploy full 300 kW AI capacity without grid infrastructure upgrades
  • Battery packs absorb peak loads and smooth power delivery across 128–144 GPUs
  • Heat recovery converts waste into usable building heat — a revenue stream
  • PUE ~1.05 — near-perfect energy efficiency at scale
~5×
Lower CAPEX vs DCs
1.05
Power Usage Effectiveness
Learn More
BATTERY AMPLIFICATION — POWER FLOW GRID IN150 kW base BATTERYLFP Storage85% OUTPUT300 kW peak 2× AMPLIFIED 24H LOAD PROFILE — 300 kW TONOFORGE™ Grid limit GPU DEMAND BATTERY + GRID 00:0006:0012:0018:00 SHAVED NO GRID UPGRADE REQUIRED · PEAK SHAVING · SMOOTH DELIVERY WASTE HEAT → RECOVERED → ADJACENT BUILDINGSCIRCULAR ENERGY · ZERO WASTE GPU TDP: MI355X 1,400W · B300 1,400W — Battery absorbs transient spikes to 1.4×
Feature 02
Rapid Deployment

Unlike traditional data centres that take 3 to 5 years to plan, permit, and build, TonoForge™ is factory pre-integrated and tested — arriving at site ready for commissioning in just 5 to 20 weeks.

Both AMD MI355X and NVIDIA GB300 NVL72 racks are pre-configured by MiTAC, Supermicro, or Dell, then integrated into the TonoForge™ container with battery, cooling, and TonoFabric™ orchestration before shipping.

  • Factory pre-integrated with liquid cooling, power, and networking
  • 5 to 20 weeks vs 3–5 years for a traditional data centre
  • Modular expansion — 150 kW to 300 kW+ without downtime
  • Re-uses existing telecom sites — no greenfield land cost
5–20w
Deployment Timeline
5 yrs
Traditional Data Centre
Learn More
DEPLOYMENT COMPARISON — TIMELINE TRADITIONAL DATA CENTRE SITE SELECTPERMITTINGCONSTRUCTIONCOMMISSION TONOFORGE™ 5–20 WKS→ LIVE TONOFORGE™ DEPLOYMENT PHASES 01 CONFIGUREGPU + Rack Select1–2 weeks 02 INTEGRATEFactory Build + DLC4–12 weeks 03 DELIVERShip + Install1–4 weeks 04 LIVECommission1–2 weeks MODULAR EXPANSION — ZERO DOWNTIME 150 kW 300 kW 600 kW n MW RE-USE EXISTING TELECOM SITES · NO GREENFIELD LAND COST RACK VENDORS: MiTAC MR1100L · Supermicro AS-4126GS / SRS-NVL72 · Dell XE9712 CHOOSE AMD MI355X (128 GPUs) OR NVIDIA GB300 NVL72 (144 GPUs) PER TONOFORGE™
Feature 03
Cost Advantage — 5x Lower

TonoForge achieves a 5x cost reduction vs traditional data centres through distributed architecture, renewable energy co-location, battery amplification, heat recovery monetisation, and re-use of existing telecom infrastructure.

Operating costs are further reduced through TonoFabric’s reinforcement learning engine, delivering 15%+ improvements in efficiency over time.

  • Co-locate with renewable energy sources
  • Re-use existing telecom sites — no greenfield land cost
  • Monetise waste heat to adjacent buildings
  • RL-based orchestration cuts operational energy costs 15%+
Lower CostHigher ResilienceSustainable ScaleSovereign AI
-5x
CAPEX vs Data Centres
15%+
RL Efficiency Gain
Learn More
CAPEX COMPARISON PER MW DEPLOYED Traditional DC EUR 10-15M / MW TonoForge EUR 2.5M 5x LESS COST SAVINGS BREAKDOWN No greenfield landReuse telecom sites No grid upgradeBattery amplification Heat monetisationRevenue from waste heat RL optimisation15%+ OpEx savings Renewable co-locationLower energy costs Modular scalingPay-as-you-grow DISTRIBUTED – RENEWABLE – CIRCULAR ECONOMY – SOVEREIGN GB300 NVL72 RACK: ~USD 3.7M | MI355X 64-GPU RACK: CONTACT VENDOR TONOFORGE COMPLETE SOLUTION: COMPUTE + POWER + COOLING + SOFTWARE
Configuration Comparison

TonoForge Configurations at a Glance

SpecificationAMD MI355X ConfigNVIDIA GB300 Config
GPU ModelAMD Instinct MI355X (CDNA 4)NVIDIA B300 Blackwell Ultra
GPUs per Rack64 (8 servers x 8 GPUs)72 (18 trays x 4 GPUs)
Total GPUs (2 Racks)128144
Memory per GPU288 GB HBM3E288 GB HBM3e
Total HBM Memory36.8 TB42 TB (21 TB/rack)
Memory Bandwidth8 TB/s per GPU8 TB/s per GPU
FP4 Peak (per rack)80.5 PFLOPS (MXFP4)1.1 EXAFLOPS (dense)
FP8 Peak (per rack)40.3 PFLOPS~360 PFLOPS
TDP per GPU1,400W1,400W
GPU Interconnect4th Gen Infinity Fabric + PCIe 5.0NVLink 5 (1.8 TB/s per GPU)
Network per GPUPensando Pollara 400 + P2200GConnectX-8 (800 Gb/s)
Host CPU2x AMD EPYC 9755 per server36x Grace CPU per rack (LPDDR5X)
System RAMUp to 6 TB DDR5-6400 / server17 TB LPDDR5X + 144x E1.S
CoolingDirect Liquid Cooling (cold plate)Full Liquid Cooling (rack-scale)
Server Form Factor4U per server (MiTAC G4826Z5)Compute tray (4 GPUs each)
Rack Format48U EIA48U rack-scale
Rack Power~120 kW per rack132 kW (8× 33kW shelf)
Rack VendorsMiTAC MR1100L-64355X-02, Supermicro AS-4126GS-TNMRSupermicro SRS-GB300-NVL72, Dell XE9712
Datatype SupportFP64/32/16 BF16 FP8 MXFP6 MXFP4FP64/32/16 BF16 FP8 FP4 INT8
Software StackROCm (PyTorch, vLLM, SGLang)CUDA (TensorRT, Triton, NeMo)

Both configurations are integrated into the TonoForge container with battery amplification, heat recovery, and TonoFabric orchestration. Power, cooling, and networking are pre-configured at factory.

Industry Benchmark

How TonoForge™ compares

Five approaches to AI infrastructure — from prefab pods to hyperscale campuses. Not all are created equal when speed, sovereignty, and sustainability converge.

Dimension APrefab Pod BPFM CoolChip CRuggedised MDC DHyperscale Campus TonoForge™Distributed AI Factory
ArchitecturePrefab pod shell
8–42 racks per pod
Modular building units
Multi-MW assembly
Shipping container
MW-class relocatable
Purpose-built campus
Multi-GW portfolio
Self-contained container
2-rack AI factory · 300 kW
Compute IntegrationNone — shell only
BYO servers & GPUs
None — shell only
BYO compute platform
GPU-ready
Configurable density
GPU clusters available
NVIDIA Preferred Partner
Factory-integrated
MI355X / GB300 NVL72
GPU Density per UnitN/A
Infrastructure only
N/A
Infrastructure only
Not disclosed
Chip-agnostic
20k+ GPUs / 50 MW
Air + liquid halls
128–144 GPUs / 300 kW
480 GPUs/MW density
Time to Deploy30% faster than on-site
Weeks to months
50% faster than on-site
Months
Weeks
Rapid relocatable
12–24+ months
Fixed campus build
5–20 weeks
Turnkey containerised
Site RequirementsPrepared building shell
Grid + permits + land
Prepared site or retrofit
Grid + permits + land
Minimal — pad or surface
Power-agnostic
1,800+ acres owned
Grid + substation + permits
Flat surface only
Existing telecom sites OK
Cooling TechnologyAir / InRow / RDHx / DLC
Flexible options
Direct-to-chip liquid
CoolChip CDU
Configurable
Air or liquid per spec
Air + liquid (new halls)
10 MW DLC facility
Full DLC integrated
250 kW CDU · N+1 pumps
PUE~1.3–1.4
Improved vs traditional
Improved vs traditional
No figure published
Not disclosed
1.1
Renewable grid
~1.05
DLC + heat recovery
Energy StrategyGrid-dependent
No on-site generation
Grid-dependent
UPS/battery backup only
Power-agnostic
Stranded gas / solar / grid
100% renewable
Grid + RECs
Battery amplification
2× factor · renewable co-loc
Heat RecoveryNot integrated
Not integrated
Not integrated
Not integrated
Circular architecture
HVAC / district heating
AI Orchestration SoftwareEcoStruxure DCIM
Facility monitoring
Vertiv DCIM
Facility monitoring
Edge Platform (AEP)
Bridge GPU orchestration
Cloud dashboard
GPU cluster mgmt
TonoFabric™
RL optimisation · workload sched.
RelocatabilityNo
Fixed installation
No
Assembled on-site
Yes
Truck / ship / rail
No
Fixed campus
Yes
ISO container · any site
Modular ScalingPod-by-pod
8–12 rack increments
Building-by-building
MW increments
Container-by-container
Leviathan stacking
Hall-by-hall
Multi-MW phases
150 kW → 300 kW → n MW
Add units to mesh
Data SovereigntyCustomer-managed
Depends on location
Customer-managed
Depends on location
Sovereign by design
US / allied nations
Fixed jurisdictions
US & Canada only
Deploy anywhere
EU sovereign · GDPR-ready
CAPEX Efficiency€10–15M / MW
Shell only — no compute
€12–16M / MW
Shell only — no compute
€12–15M / MW
Undisclosed detail
€10–15M / MW
Fixed campus scale
~€2.5M / MW
5–6× less · compute included
Best Suited ForEnterprise white-space fit-out inside existing facilitiesNew-build or retrofit AI halls at prepared sitesRemote / defence edge AI in contested environmentsHyperscale cloud & GPU-as-a-service at fixed campusesDistributed sovereign AI at telecom / renewable sites

Benchmark based on publicly available product specifications and press materials as of Q3 2025. A–D represent representative industry approaches, not specific endorsements or claims about named competitors. TonoForge™ figures reflect Tonomia design specifications.

TRADITIONAL DATA CENTRE HVAC UPS PDU GENSET CRAH COOLING TOWERSHEAT WASTED HV GRIDREQUIRED 10-15M/MWCapital cost 3-5 YEARSBuild time PUE 1.4-1.8Energy waste CENTRALIZED – GRID-DEPENDENT – HIGH-COST – SLOW 120+ kW per rack requires HV grid connection + cooling infrastructure
TONOFORGE DISTRIBUTED MODULE GPU COMPUTE BATTERY COOLING POWER 800G NET SECURITY ALL-IN-ONE PRE-INTEGRATED SOLAR FIBER HEAT BUILDINGS 2.5M/MWCapital cost 5-20 WEEKSDeploy time PUE ~1.05Near-perfect DISTRIBUTED – GRID-FREE – LOW-COST – FAST 2x RACKS: MI355X (128 GPU) or GB300 NVL72 (144 GPU)
DISTRIBUTED NETWORK MESH TONOFORGE DEPLOYMENT TOPOLOGY TF-01MI355X TF-02GB300 FABRICCORE TF-03MI355X TF-04GB300 EDGE EDGE EDGE EDGE TONOFORGE (CORE) TONOEDGE (EDGE) TONOFABRIC (ORCHESTRATION) HETEROGENEOUS GPU MESH
AI at the Core Network

Deploy TonoForge at your core network

Connect with our engineering team to explore a deployment tailored to your GPU requirements, network topology, energy profile, and scale. Choose AMD MI355X or NVIDIA GB300 NVL72 from MiTAC, Supermicro, or Dell.