Stacks
Configure your AI infrastructure with model selection, hardware preferences, and framework choices. Get AI-suggested stacks optimized for your scenario.

Model Selection
Choose from supported LLM providers and models. Compare capabilities, pricing, and performance characteristics to find the right fit for your use case.
- Anthropic Claude (Haiku, Sonnet, Opus)
- OpenAI GPT-4, GPT-4o, o1
- Google Gemini, Meta Llama, Mistral
Hardware Preferences
Specify GPU requirements for self-hosted deployments. Lattice helps you understand hardware needs for different models and throughput targets.
- NVIDIA GPU options (A100, H100, L40S)
- Memory requirements per model
- Cloud vs. on-premises deployment
Framework Configuration
Select inference frameworks and serving solutions. From vLLM to TensorRT-LLM, choose the right tools for your deployment requirements.
- vLLM for high-throughput serving
- TensorRT-LLM for optimized inference
- Triton Inference Server integration
AI-Suggested Stacks
Let Lattice recommend optimal stack configurations based on your scenario. Get suggestions grounded in vendor documentation and real-world benchmarks.
- Scenario-optimized recommendations
- Cost-performance tradeoff analysis
- Citations to vendor benchmarks
Configure Your AI Stack
Get AI-powered stack recommendations optimized for your specific requirements.
Get Lattice for $99