Arcee AI Releases Trinity Large Thinking: Free 262K Context Reasoning Model
Arcee AI has released Trinity Large Thinking, an open source reasoning model with a 262,144-token context window. The model is available free via OpenRouter and claims strong performance in PinchBench, agentic workloads, and reasoning tasks.
Trinity Large Thinking — Quick Specs
Arcee AI Releases Trinity Large Thinking: Free 262K Context Reasoning Model
Arcee AI has released Trinity Large Thinking, an open source reasoning model with a 262,144-token context window, available free via OpenRouter as of April 1, 2026.
Key Specifications
- Context window: 262,144 tokens
- Pricing: $0 per million input tokens, $0 per million output tokens
- Model type: Reasoning model with step-by-step thinking capabilities
- Availability: Free tier on OpenRouter platform
Performance Claims
According to Arcee AI, Trinity Large Thinking shows strong performance across:
- PinchBench evaluations
- Agentic workloads
- Reasoning tasks
Specific benchmark scores have not been disclosed at launch.
Technical Implementation
The model supports OpenRouter's reasoning-enabled API, which allows developers to access the model's internal reasoning process through the reasoning_details array in API responses. Developers can use the reasoning parameter to enable step-by-step thinking output.
When continuing conversations, the complete reasoning_details must be preserved in message history for the model to maintain reasoning continuity.
Open Source Availability
Model weights are available for download, following Arcee AI's open source approach. The company has published a launch video detailing the model's capabilities and use cases.
What This Means
Trinity Large Thinking enters a competitive reasoning model market currently dominated by OpenAI's o1 and o3 series, Anthropic's extended thinking modes, and DeepSeek's R1. The 262K context window positions it among the larger-context reasoning models available, though it's unclear how this compares to specialized long-context models.
The free tier availability via OpenRouter makes it immediately accessible for developers to test against commercial alternatives. However, without published benchmark scores on standard reasoning evaluations like AIME or GPQA, performance relative to established models remains unverified. The emphasis on "agentic workloads" suggests optimization for multi-step tool use and planning tasks rather than pure mathematical reasoning.
Related Articles
Baidu Launches CoBuddy Code Generation Model with 131K Context Window, Free on OpenRouter
Baidu has released CoBuddy, a code generation model optimized for coding tasks and AI agent workflows. The model features a 131K token context window, up to 65K output tokens, and runs on fp8 quantization with native support for tool calling and reasoning.
IBM Releases Granite Speech 4.1 2B: 2-Billion-Parameter Multilingual Speech Model with Non-Autoregressive Variant
IBM has released Granite Speech 4.1 2B, a 2-billion-parameter speech-language model trained on 174,000 hours of audio for automatic speech recognition and translation across English, French, German, Spanish, Portuguese, and Japanese. The model introduces a dual-head CTC encoder and includes variants for speaker attribution and a novel non-autoregressive architecture for higher throughput.
OpenAI offers EU preview access to GPT-5.5-Cyber model while Anthropic withholds Mythos
OpenAI announced GPT-5.5-Cyber is rolling out in limited preview to vetted cybersecurity teams and is in discussions with the European Commission about preview access. Anthropic released its Mythos model a month ago but has yet to grant EU access for security review.
Supertone releases Supertonic 3: 99M-parameter on-device TTS model supporting 31 languages
Supertone has released Supertonic 3, a 99M-parameter text-to-speech model that runs entirely on-device using ONNX Runtime. The model expands language support from 5 to 31 languages compared to Supertonic 2, requires no GPU, and claims competitive accuracy against models 7-20x larger.
Comments
Loading...