Stacks

Configure your AI infrastructure with model selection, hardware preferences, and framework choices. Get AI-suggested stacks optimized for your scenario.

latticelab.io
Lattice Stack editor showing model provider selection, temperature, and framework configuration

Model Selection

Choose from supported LLM providers and models. Compare capabilities, pricing, and performance characteristics to find the right fit for your use case.

  • Anthropic Claude (Haiku, Sonnet, Opus)
  • OpenAI GPT-4, GPT-4o, o1
  • Google Gemini, Meta Llama, Mistral

Hardware Preferences

Specify GPU requirements for self-hosted deployments. Lattice helps you understand hardware needs for different models and throughput targets.

  • NVIDIA GPU options (A100, H100, L40S)
  • Memory requirements per model
  • Cloud vs. on-premises deployment

Framework Configuration

Select inference frameworks and serving solutions. From vLLM to TensorRT-LLM, choose the right tools for your deployment requirements.

  • vLLM for high-throughput serving
  • TensorRT-LLM for optimized inference
  • Triton Inference Server integration

AI-Suggested Stacks

Let Lattice recommend optimal stack configurations based on your scenario. Get suggestions grounded in vendor documentation and real-world benchmarks.

  • Scenario-optimized recommendations
  • Cost-performance tradeoff analysis
  • Citations to vendor benchmarks

Configure Your AI Stack

Get AI-powered stack recommendations optimized for your specific requirements.

Get Lattice for $99