Skip to content
Prev Previous commit
added vllm to docs
  • Loading branch information
waleedlatif1 committed Nov 26, 2025
commit d0e639e302120c39f7456075931c827f3778e871
2 changes: 1 addition & 1 deletion apps/docs/content/docs/en/blocks/agent.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -46,7 +46,7 @@ The Agent block supports multiple LLM providers through a unified inference inte
- **Anthropic**: Claude 4.5 Sonnet, Claude Opus 4.1
- **Google**: Gemini 2.5 Pro, Gemini 2.0 Flash
- **Other Providers**: Groq, Cerebras, xAI, Azure OpenAI, OpenRouter
- **Local Models**: Ollama-compatible models
- **Local Models**: Ollama or VLLM compatible models

### Temperature

Expand Down
2 changes: 1 addition & 1 deletion apps/docs/content/docs/en/blocks/evaluator.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -52,7 +52,7 @@ Choose an AI model to perform the evaluation:
- **Anthropic**: Claude 3.7 Sonnet
- **Google**: Gemini 2.5 Pro, Gemini 2.0 Flash
- **Other Providers**: Groq, Cerebras, xAI, DeepSeek
- **Local Models**: Ollama-compatible models
- **Local Models**: Ollama or VLLM compatible models

Use models with strong reasoning capabilities like GPT-4o or Claude 3.7 Sonnet for best results.

Expand Down
2 changes: 1 addition & 1 deletion apps/docs/content/docs/en/blocks/guardrails.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -65,7 +65,7 @@ Uses Retrieval-Augmented Generation (RAG) with LLM scoring to detect when AI-gen
**Configuration:**
- **Knowledge Base**: Select from your existing knowledge bases
- **Model**: Choose LLM for scoring (requires strong reasoning - GPT-4o, Claude 3.7 Sonnet recommended)
- **API Key**: Authentication for selected LLM provider (auto-hidden for hosted/Ollama models)
- **API Key**: Authentication for selected LLM provider (auto-hidden for hosted/Ollama or VLLM compatible models)
- **Confidence Threshold**: Minimum score to pass (0-10, default: 3)
- **Top K** (Advanced): Number of knowledge base chunks to retrieve (default: 10)

Expand Down
2 changes: 1 addition & 1 deletion apps/docs/content/docs/en/blocks/router.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -56,7 +56,7 @@ Choose an AI model to power the routing decision:
- **Anthropic**: Claude 3.7 Sonnet
- **Google**: Gemini 2.5 Pro, Gemini 2.0 Flash
- **Other Providers**: Groq, Cerebras, xAI, DeepSeek
- **Local Models**: Ollama-compatible models
- **Local Models**: Ollama or VLLM compatible models

Use models with strong reasoning capabilities like GPT-4o or Claude 3.7 Sonnet for best results.

Expand Down
3 changes: 2 additions & 1 deletion apps/docs/content/docs/en/execution/costs.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -73,6 +73,7 @@ The model breakdown shows:
| Groq | Llama 4 Scout | $0.40 / $0.60 |
| Cerebras | Llama 3.3 70B | $0.94 / $0.94 |
| Ollama | Local models | Free |
| VLLM | Local models | Free |

*Pay providers directly with no markup*
</Tab>
Expand All @@ -86,7 +87,7 @@ The model breakdown shows:

- **Model Selection**: Choose models based on task complexity. Simple tasks can use GPT-4.1-nano while complex reasoning might need o1 or Claude Opus.
- **Prompt Engineering**: Well-structured, concise prompts reduce token usage without sacrificing quality.
- **Local Models**: Use Ollama for non-critical tasks to eliminate API costs entirely.
- **Local Models**: Use Ollama or VLLM for non-critical tasks to eliminate API costs entirely.
- **Caching and Reuse**: Store frequently used results in variables or files to avoid repeated AI model calls.
- **Batch Processing**: Process multiple items in a single AI request rather than making individual calls.

Expand Down
2 changes: 1 addition & 1 deletion apps/docs/content/docs/en/introduction/index.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -59,7 +59,7 @@ Enable your team to build together. Multiple users can edit workflows simultaneo

Sim provides native integrations with 80+ services across multiple categories:

- **AI Models**: OpenAI, Anthropic, Google Gemini, Groq, Cerebras, local models via Ollama
- **AI Models**: OpenAI, Anthropic, Google Gemini, Groq, Cerebras, local models via Ollama or VLLM
- **Communication**: Gmail, Slack, Microsoft Teams, Telegram, WhatsApp
- **Productivity**: Notion, Google Workspace, Airtable, Monday.com
- **Development**: GitHub, Jira, Linear, automated browser testing
Expand Down
1 change: 0 additions & 1 deletion apps/sim/tools/incidentio/escalations_create.ts
Original file line number Diff line number Diff line change
Expand Up @@ -66,7 +66,6 @@ export const escalationsCreateTool: ToolConfig<
}

if (params.user_ids) {
// Split comma-separated string into array
body.user_ids = params.user_ids.split(',').map((id: string) => id.trim())
}

Expand Down