Seclai

Seclai

Pricing

Qwen3 235B A22B Instruct

LLM
Qwen

Qwen3 235B A22B Instruct excels at instruction following, reasoning, coding, and tool use with a 256K context window, multilingual support across 100+ languages, and strong long-context understanding.

Back to models

Context tokens

262,144

Output tokens

8,192

Released

Apr 29, 2025

Schema

Qwen uses the OpenAI API schema for message creation. We handle the model, stream, service_tier, safety_identifier, and user parameters.

Schema documentation

Capabilities

Tool use
Structured output
Thinking
OpenAI API
Multilingual

Supported languages

ar
de
en
es
fr
it
ja
ko
pt
zh

Supported tools

  • Seclai Content Tools

    Inspect source documents connected to your account. Includes tools for loading full content, reading character ranges, searching within documents, viewing stats, and listing available content sources. When a source_connection_content_version_id is provided in agent run metadata it is used as the default. Otherwise the model can discover content via list_content_sources.

  • Seclai Knowledge Base

    Search your knowledge bases using semantic similarity. Includes search_knowledge_base and list_knowledge_bases. When a knowledge_base_id is provided in the prompt or agent run metadata it is used as the default. Otherwise the model can discover available knowledge bases at runtime.

Variants

Tier

Priority and flex tiers trade speed for cost.

OptionDescriptionInput credits (per 1k tokens)Output credits (per 1k tokens)
FlexFlex processing provides lower costs for Responses or Chat Completions requests in exchange for slower response times and occasional resource unavailability. It's ideal for non-production or lower priority tasks, such as model evaluations, data enrichment, and asynchronous workloads.1.465.85
PriorityPriority processing delivers significantly lower and more consistent latency compared to Standard processing while keeping pay-as-you-go flexibility. Priority processing is ideal for high-value, user-facing applications with regular traffic where latency is paramount. Priority processing should not be used for data processing, evaluations, or other highly erratic traffic.5.1220.48
Standard2.9311.70