PLATFORM // GPU COMPUTE

MANAGED GPU INFRASTRUCTURE
FOR VLM TRAINING.

From T4 to B200, select the right GPU tier for your workload. Multi-GPU scaling with NVLink, VRAM-aware scheduling, and background job queues. No DevOps required.

Datature/Hardware/GPU Configuration
● Available
GPU TypeVRAMCUDA CoresBest ForMulti-GPUTier
NVIDIAT4
16 GB2,560Inference, Small LoRAUp to 4Starter
NVIDIAL4
24 GB7,424LoRA Fine-TuningUp to 8Developer
NVIDIAA10
24 GB9,216General Purpose TrainingUp to 8Developer
NVIDIAA100Recommended
80 GB6,912Production SFT TrainingUp to 32Developer
NVIDIAH100
80 GB16,896Large-Scale Training, NVLinkUp to 64Professional
NVIDIAB200
192 GB18,000+Largest Models, Multi-NodeUp to 64Enterprise
Estimated VRAM UsageA100 80GB Selected
Required: 56GBAvailable: 80GB

MULTI-GPU SCALING

SCALE FROM 1 TO 64 GPUS.

Vi manages multi-GPU orchestration automatically. Select the number of GPUs in the hardware configuration modal and Vi handles data parallelism, gradient synchronization, and NVLink topology.

Up to 64GPUs Per Run

High-Bandwidth Interconnect

Multi-GPU runs on H100 and B200 tiers use NVLink for direct GPU-to-GPU communication. Vi configures the topology automatically.

AutomaticParallelism Config

Distributed Training

Vi handles parallelism strategy selection based on your model size and GPU count. No manual configuration of sharding or gradient sync required.

Per-EpochCheckpoint Frequency

Automatic Checkpointing

Model checkpoints saved periodically during training. Resume from any checkpoint if a run is interrupted or you want to branch from an earlier state.

AsyncExecution Mode

Background Execution

Training runs server-side on managed clusters. Close your browser, shut your laptop. Vi notifies you when the run completes.

Pre-FlightValidation Check

VRAM-Aware Scheduling

Vi estimates VRAM requirements before provisioning hardware. If your config exceeds the selected GPU tier, Vi warns you before launch.

Real-TimeDashboard Updates

Run Monitoring

Track loss curves, epoch progress, and GPU utilization in real time from the dashboard. Email and in-app notifications on completion.

VRAM ESTIMATOR

KNOW BEFORE YOU TRAIN.

Vi estimates VRAM requirements based on your model size, training method, and batch configuration. Select a model and method to see the estimated VRAM and the recommended GPU tier.

  • Model Parameters

    The base model parameter count is the primary driver of VRAM usage. A 7B model requires roughly 14 GB in FP16 just for weights alone.

  • Training Method

    LoRA adapters add minimal overhead (typically 1-5% of base parameters). Full SFT requires 2-3x the model weight size for optimizer states and gradients.

  • Batch Size

    Each sample in the batch requires activation memory. Gradient checkpointing trades compute for memory, reducing activation overhead by 60-80%.

  • Sequence Length

    Longer input sequences require quadratically more attention memory. Flash Attention 2 reduces this to near-linear scaling.

VRAM Estimation Table

Qwen 7B

LoRA (FP16)

12 GB

Qwen 7B

Full SFT (FP16)

28 GB

Qwen 7B

LoRA (NF4)

7 GB

Qwen 32B

LoRA (FP16)

48 GB

Qwen 32B

Full SFT (FP16)

128 GB

Qwen 72B

LoRA (NF4)

36 GB

InternVL 8B

LoRA (FP16)

14 GB

InternVL 38B

Full SFT (FP16)

142 GB

Estimates include model weights, optimizer states, gradients, and activation memory with gradient checkpointing enabled. Actual usage may vary by 10-15% depending on sequence length and batch size.

INFRASTRUCTURE

YOUR GPUS OR OURS. VI MANAGES BOTH.

Use Vi Cloud for zero-setup managed training, or connect your own GPU cluster with custom runners. Your data stays in your infrastructure. Vi orchestrates the training pipeline either way.

Your Data

S3 / Azure / GCS

Base Model

Qwen / InternVL / Cosmos

Your GPUs

Custom Runners (BYOG)

Vi

Datature Vi

Orchestration Layer

Annotate / Train / EvaluateDPO / Deploy / Monitor

Vi Cloud

T4 to B200 Managed

Deployment

SDK / NIM / API

Monitoring

Metrics / Alerts / Logs

Vi Cloud (Managed)

Zero-setup GPU access from T4 to B200. Vi provisions, configures, and deallocates hardware automatically. Pay per training run.

Bring Your Own GPU

Connect your existing GPU cluster via custom runners. Vi orchestrates training on your hardware while keeping data in your infrastructure.

Hybrid

Use Vi Cloud for development and prototyping. Switch to your own cluster for production runs. Same training config, different compute backend.

COST MANAGEMENT

TRAIN MORE. SPEND LESS.

Vi routes training jobs to the most cost-effective instance type for your workload. Background job queues batch runs for off-peak scheduling, the VRAM estimator prevents over-provisioning, and automatic instance selection keeps costs low.

Smart

Instance Routing

Vi routes training jobs to the most cost-effective instance type for your workload. Spot instances when available, on-demand when needed. Automatic failover between instance types.

Queue

Background Job Queue

Launch training runs and close your browser. Vi executes jobs in the background and sends email notifications when runs complete or encounter errors.

Auto

Right-Size GPU Selection

The VRAM estimator recommends the cheapest GPU tier that fits your workload. Stop paying for A100s when a T4 is sufficient for your LoRA fine-tune.

EDU

Academic Pricing

Discounted compute rates for verified academic and research institutions. Contact the Vi team with your institutional email for eligibility verification.

Vi Cloud (Managed)

Fully managed GPU infrastructure. No cloud account required. Vi provisions, schedules, and deallocates GPUs automatically. Pay per GPU-hour with per-second billing.

T4L4A10A100H100B200

Bring Your Own Cloud

Enterprise

Connect your existing AWS, GCP, or Azure account. Vi orchestrates training jobs on your cloud infrastructure while you retain full control over billing and data residency.

AWSGoogle CloudAzure

WORKFLOW

THREE STEPS. ZERO DEVOPS.

Configuring and launching a GPU training run takes under two minutes. No Dockerfiles, no CUDA drivers, no Kubernetes manifests. Select your hardware, configure the run, and monitor results.

01

Select GPU

Open the Hardware Configuration modal. Choose your GPU type and quantity from the available tiers. The VRAM estimator shows whether your selection fits the model and training method.

  • 6 GPU tiers from T4 to B200
  • VRAM usage bar updates in real-time
  • Multi-GPU slider from 1 to 64
  • Spot instance toggle for cost savings
02

Configure Run

Set your training hyperparameters, dataset split, and output checkpoint location. Vi validates the full configuration against the selected hardware before launch.

  • Learning rate, epochs, batch size
  • LoRA rank and target modules
  • Quantization mode (FP16 / INT8 / NF4)
  • Checkpoint save interval and storage path
03

Monitor and Results

Launch the run and track progress in real-time. Loss curves, VRAM usage, and throughput metrics stream to the Neural Monitor dashboard. Email notification on completion.

  • Live loss and metric curves
  • GPU utilization and temperature
  • Automatic checkpoint saving
  • Email notification on completion
training-config.json

{

"gpu_type": "A100",

"gpu_count": 4,

"model": "Qwen2.5-VL-7B",

"method": "lora",

"quantization": "fp16",

"lora_rank": 16,

"epochs": 3,

"batch_size": 8,

"learning_rate": 2e-4,

"spot_instance": true,

"checkpoint_interval": 500,

"notify_on_complete": true

}

STOP CONFIGURING INFRA.
START TRAINING MODELS.

Managed GPU infrastructure from T4 to B200. Start free today.