Skip to content

Models & Aliases

llmist supports multiple LLM providers with convenient aliases for quick access.

Use short aliases instead of full model names:

AliasFull Model NameProvider
gpt5gpt-5OpenAI
gpt5-minigpt-5-miniOpenAI
gpt4ogpt-4oOpenAI
o4-minio4-miniOpenAI
sonnetclaude-sonnet-4-5Anthropic
opusclaude-opus-4-5Anthropic
haikuclaude-haiku-4-5Anthropic
flashgemini-2.5-flashGoogle
progemini-3-pro-previewGoogle

For explicit provider selection, use the provider:model format:

Terminal window
# Explicit provider selection
npx @llmist/cli complete "Hello" --model openai:gpt-5
npx @llmist/cli complete "Hello" --model anthropic:claude-sonnet-4-5
npx @llmist/cli complete "Hello" --model gemini:gemini-2.5-flash
npx @llmist/cli complete "Hello" --model huggingface:meta-llama/Llama-3.1-8B-Instruct
ModelVisionStreamingTool UseReasoningContext
GPT-5128K
GPT-5 Mini128K
GPT-4o128K
o3200K
o4-mini200K
Claude Opus 4.5200K
Claude Sonnet 4.5200K
Claude Haiku 4.5200K
Gemini Flash1M
Gemini Pro1M
DeepSeek V3.264K
DeepSeek R164K
Llama 3.3 70B128K
Qwen 2.5 72B128K
Mixtral 8x7B32K
Use CaseRecommendedWhy
General taskssonnetBest balance of quality and speed
Complex reasoningopus, o3, DeepSeek R1Highest capability; use .withReasoning("high") for o3/Claude
High-volume taskshaiku, flashFast and cost-effective
Long documentsflash, pro1M token context
Codingsonnet, gpt5, DeepSeek V3.2Strong code understanding
Vision tasksgpt4o, flashExcellent image analysis
Open-source/self-hostedLlama 3.3 70B, Qwen 2.5 72BFree via HuggingFace

llmist supports popular open-source models via HuggingFace’s serverless inference:

  • deepseek-ai/DeepSeek-V3.2 - 685B MoE model for general reasoning and tool use
  • deepseek-ai/DeepSeek-R1 - Reasoning model excelling at math, logic, and coding
  • deepseek-ai/DeepSeek-Coder-V2-Instruct - Specialized for code generation
  • meta-llama/Llama-3.3-70B-Instruct - General-purpose flagship model
  • meta-llama/Llama-3.1-8B-Instruct - Efficient smaller variant
  • meta-llama/Llama-3.2-11B-Vision-Instruct - Vision-enabled model
  • Qwen/Qwen2.5-72B-Instruct - Strong general-purpose model
  • Qwen/Qwen2.5-Coder-32B-Instruct - Code-specialized
  • Qwen/Qwen2-VL-72B-Instruct - Vision-language model
  • mistralai/Mixtral-8x7B-Instruct-v0.1 - Mixture-of-experts architecture
  • mistralai/Mistral-Nemo-Instruct-2407 - 12B efficient model
ModelProviderDescription
dall-e-3OpenAIHigh-quality image generation
dall-e-2OpenAIFaster, lower cost
imagen-3GoogleGemini image generation
ModelProviderDescription
tts-1OpenAIText-to-speech, standard quality
tts-1-hdOpenAIText-to-speech, high quality
import { LLMist } from 'llmist';
// Using alias
const answer = await LLMist.createAgent()
.withModel('sonnet')
.askAndCollect('Hello!');
// Using full name
const answer2 = await LLMist.createAgent()
.withModel('anthropic:claude-sonnet-4-5')
.askAndCollect('Hello!');
// HuggingFace model
const answer3 = await LLMist.createAgent()
.withModel('huggingface:meta-llama/Llama-3.1-8B-Instruct')
.askAndCollect('Hello!');
// HuggingFace with provider routing
const answer4 = await LLMist.createAgent()
.withModel('hf:deepseek-ai/DeepSeek-V3.2:fastest')
.askAndCollect('Write a Python function to calculate factorial');
// OpenRouter model (access 400+ models)
const answer5 = await LLMist.createAgent()
.withModel('openrouter:deepseek/deepseek-chat')
.askAndCollect('Explain quantum computing');
// OpenRouter with alias
const answer6 = await LLMist.createAgent()
.withModel('or:sonnet')
.askAndCollect('Hello!');

llmist automatically discovers available providers based on environment variables:

VariableProvider
OPENAI_API_KEYOpenAI
ANTHROPIC_API_KEYAnthropic
GEMINI_API_KEYGoogle Gemini
OPENROUTER_API_KEYOpenRouter
HF_TOKENHuggingFace

See Environment Variables for complete configuration.