Skip to content

Models Reference

Sciorex supports multiple AI providers, giving you flexibility to choose the best model for your use case.

Supported Providers

ProviderTypeBest For
Claude CodeCloudComplex reasoning, coding, analysis
Google GeminiCloudFast responses, multimodal, large context
OpenAI CodexCloudCode generation, general tasks
LM StudioLocalPrivacy, offline, experimentation
OllamaLocalPrivacy, custom models, self-hosted

Setup Instructions

See Settings → AI Providers in the app for setup instructions for each provider.

Claude Models

Claude models excel at complex reasoning, coding tasks, and nuanced understanding.

ModelIDCapabilitiesContext
Claude Opus 4.5claude-opus-4-5-20251101Reasoning, Vision, Tools, PDF, Code200K
Claude Sonnet 4.5claude-sonnet-4-5-20250929Fast, Reasoning, Vision, Tools, PDF, Code200K
Claude Haiku 4.5claude-haiku-4-5-20251001Fast, Vision, Tools, Code200K
Claude Opus 4.1claude-opus-4-1-20250805Reasoning, Vision, Tools, PDF, Code200K
Claude Opus 4claude-opus-4-20250514Reasoning, Vision, Tools, PDF, Code200K
Claude Sonnet 4claude-sonnet-4-20250514Reasoning, Vision, Tools, PDF, Code200K
Claude Haiku 3.5claude-3-5-haiku-20241022Fast, Vision, Tools, Code200K

Default Choice

Claude Sonnet 4.5 is the recommended default for most coding tasks.

Gemini Models

Gemini models offer fast responses and excellent multimodal capabilities.

ModelIDCapabilitiesContext
Gemini 2.5 Progemini-2.5-proReasoning, Vision, Tools, PDF, Code1M
Gemini 2.5 Flashgemini-2.5-flashFast, Vision, Tools, PDF, Code1M
Gemini 2.5 Flash Litegemini-2.5-flash-liteFast, Tools, Code128K

Large Context

Gemini 2.5 Pro and Flash support up to 1M tokens context window.

OpenAI Codex Models

OpenAI Codex models provide strong coding capabilities with large context windows.

Codex Models

ModelIDCapabilitiesContext
GPT-5.1 Codex Maxgpt-5.1-codex-maxReasoning, Tools, PDF, Code400K
GPT-5.1 Codexgpt-5.1-codexReasoning, Tools, PDF, Code400K
GPT-5.1 Codex Minigpt-5.1-codex-miniFast, Tools, Code400K
GPT-5 Codexgpt-5-codexReasoning, Tools, PDF, Code400K
GPT-5 Codex Minigpt-5-codex-miniFast, Tools, Code400K

GPT Models

ModelIDCapabilitiesContext
GPT-5.2gpt-5.2Reasoning, Vision, Tools, PDF, Code400K
GPT-5.1gpt-5.1Reasoning, Vision, Tools, PDF, Code400K
GPT-5gpt-5Reasoning, Vision, Tools, PDF, Code400K

GPT-OSS Models (Local)

These models run locally via Codex CLI with the --oss flag.

ModelIDCapabilitiesContext
GPT-OSS 120Bopenai/gpt-oss-120bReasoning, Code, Tools128K
GPT-OSS 20Bopenai/gpt-oss-20bFast, Code, Tools128K

Local Models (LM Studio & Ollama)

Run models locally for privacy, offline work, or cost savings. Models are fetched dynamically from your local server.

LM Studio

Popular models for LM Studio:

ModelParametersBest ForVRAM Required
Qwen 2.5 Coder7B / 32BCode generation8GB / 24GB
DeepSeek Coder V216BAdvanced coding12GB
CodeLlama7B / 34BCode completion8GB / 24GB
Llama 3.23B / 11BGeneral tasks4GB / 12GB

Ollama

Popular models for Ollama:

ModelBest For
qwen2.5-coderCode generation
deepseek-coder-v2Advanced coding
codellamaCode tasks
llama3.2General tasks
mistralFast responses

Hardware Requirements

  • 7B models: 8GB+ VRAM
  • 13B models: 16GB+ VRAM
  • 34B+ models: 24GB+ VRAM

Model Capabilities

CapabilityDescription
FastOptimized for speed
VisionCan analyze images
ReasoningAdvanced reasoning capabilities
ToolsCan use tools and function calling
PDFCan read and analyze PDFs
CodeOptimized for code generation
Long ContextLarge context window
Image GenCan generate images

Provider Comparison

By Use Case

Use CaseRecommended ProviderRecommended Model
Complex architectureClaudeOpus 4.5
General codingClaude / GeminiSonnet 4.5 / Gemini 2.5 Flash
Quick tasksGemini / CodexGemini 2.5 Flash Lite / Codex Mini
Code reviewClaudeSonnet 4.5
Privacy-sensitiveLocalLM Studio / Ollama
Offline workLocalAny local model
Long contextGemini / CodexGemini 2.5 Pro (1M) / GPT-5.1 (400K)

By Feature

FeatureClaudeGeminiCodexLocal
Extended Thinking
Tool UseVaries
Vision/ImagesVaries
Max Context200K1M400KVaries
Offline

Extended Thinking

LevelToken BudgetClaudeGeminiCodex
Off0
Think1,024
Think Hard10,000
Think Harder16,000--
Ultrathink32,000--

Configuring in Agents

yaml
name: Code Reviewer
provider: claude-code
model: claude-sonnet-4-5-20250929
thinkingLevel: think
yaml
name: Fast Analyzer
provider: google-gemini
model: gemini-2.5-flash
yaml
name: Local Formatter
provider: ollama
model: qwen2.5-coder

Pricing

Claude

ModelInputOutput
Opus 4.5$15/1M$75/1M
Sonnet 4.5$3/1M$15/1M
Haiku 4.5$1/1M$5/1M
Haiku 3.5$0.80/1M$4/1M

Gemini

ModelInputOutput
Gemini 2.5 Pro$1.25/1M$5/1M
Gemini 2.5 Flash$0.075/1M$0.30/1M
Gemini 2.5 Flash Lite$0.019/1M$0.075/1M

Codex

ModelInputOutput
GPT-5.2$1.75/1M$14/1M
GPT-5.1 Codex Max$1.25/1M$10/1M
GPT-5.1 Codex Mini$0.25/1M$2/1M

Best Practices

For Agents

yaml
# Complex analysis
name: Senior Reviewer
provider: claude-code
model: claude-opus-4-5-20251101
thinkingLevel: think-hard

# Fast local tasks
name: Code Formatter
provider: ollama
model: qwen2.5-coder

# Quick cloud analysis
name: Quick Analyzer
provider: google-gemini
model: gemini-2.5-flash

For Flows

  • Use local models or Gemini Flash for transformation nodes
  • Use Claude Sonnet or GPT-5.1 for decision nodes
  • Use Claude Opus for critical analysis nodes

Released under the MIT License.