stacks hardware infrastructure

Configuring Stack Hardware Settings

Lattice Lab
Configuring Stack Hardware Settings

When I need to specify infrastructure for my AI workload, I want to configure hardware settings for my stack, so I can control cloud provider, GPU type, and compute resources.


Introduction

In Lattice, the Hardware section of a stack defines where and how your AI workload runs. This includes cloud provider selection (AWS, GCP, Azure), GPU type for inference or training (H100, A100, T4), instance configuration, and region selection.

Hardware configuration directly impacts cost, latency, and performance. The right configuration depends on your workload requirements, budget constraints, and geographic needs.

Step 1: Access Stack Configuration

Step 1: Lab panel with ContextBar showing current stack

The ContextBar displays your current stack:

Quick Access

  • Stack Name: Shows which stack you’re using
  • Edit Button: Pencil icon opens configuration
  • Scenario Context: Your workload requirements

Click the edit button to open the StackConfigForm and navigate to hardware settings.

Step 2: Stack Configuration Form

Step 2: StackConfigForm modal showing Model section

The StackConfigForm opens with sections for complete configuration:

Available Sections

  • Model: AI provider and model settings
  • Framework: Orchestration and observability
  • Hardware: Cloud and compute configuration (scroll to access)

Scroll down past the Framework section to reach Hardware settings.

Step 3: Cloud Provider and Instance

Step 3: Hardware section showing cloud provider and instance options

The Hardware section shows cloud configuration:

Cloud Provider

  • AWS: Amazon Web Services
  • GCP: Google Cloud Platform
  • Azure: Microsoft Azure

Instance Configuration

  • Instance Family: GPU-optimized, compute-optimized, memory-optimized
  • Instance Size: Small, medium, large, xlarge
  • Spot Instances: Cost-saving option with interruption risk

Provider selection affects available instance types, pricing, and regional availability.

Step 4: GPU and Region Configuration

Step 4: GPU type and region configuration options

Scrolling further reveals GPU and region settings:

GPU Type

  • H100: Latest generation, highest performance
  • A100: High performance, widely available
  • A10G: Balanced performance and cost
  • T4: Cost-effective for inference

Region Selection

  • Primary Region: Main deployment location
  • Latency Optimization: Choose regions close to users
  • Data Residency: Meet compliance requirements

Advanced Options

  • Multi-GPU: Configure multiple GPUs per instance
  • Reserved Capacity: Guaranteed availability
  • Spot Fallback: Automatic fallback from spot to on-demand

What You’ve Accomplished

You’ve learned how to configure hardware settings in Lattice:

  • Access configuration — Click edit button on current stack
  • Navigate to Hardware — Scroll past Model and Framework sections
  • Select cloud provider — Choose AWS, GCP, or Azure
  • Configure instances — Set family, size, and spot options
  • Choose GPU — Select appropriate GPU for your workload
  • Set region — Pick deployment region

Key benefits of hardware configuration:

  • Cost control — Choose cost-effective options for your budget
  • Performance optimization — Select GPUs matching your latency needs
  • Geographic placement — Deploy close to your users
  • Compliance — Meet data residency requirements

Next steps: Compare GPU options for your latency requirements, or evaluate spot instance savings for non-critical workloads.

Ready to Try Lattice?

Get lifetime access to Lattice for confident AI infrastructure decisions.

Get Lattice for $99