model releaseArcee Ai

Arcee AI Releases Trinity Large Thinking: Free 262K Context Reasoning Model

TL;DR

Arcee AI has released Trinity Large Thinking, an open source reasoning model with a 262,144-token context window. The model is available free via OpenRouter and claims strong performance in PinchBench, agentic workloads, and reasoning tasks.

1 min read
0

Arcee AI Releases Trinity Large Thinking: Free 262K Context Reasoning Model

Arcee AI has released Trinity Large Thinking, an open source reasoning model with a 262,144-token context window, available free via OpenRouter as of April 1, 2026.

Key Specifications

  • Context window: 262,144 tokens
  • Pricing: $0 per million input tokens, $0 per million output tokens
  • Model type: Reasoning model with step-by-step thinking capabilities
  • Availability: Free tier on OpenRouter platform

Performance Claims

According to Arcee AI, Trinity Large Thinking shows strong performance across:

  • PinchBench evaluations
  • Agentic workloads
  • Reasoning tasks

Specific benchmark scores have not been disclosed at launch.

Technical Implementation

The model supports OpenRouter's reasoning-enabled API, which allows developers to access the model's internal reasoning process through the reasoning_details array in API responses. Developers can use the reasoning parameter to enable step-by-step thinking output.

When continuing conversations, the complete reasoning_details must be preserved in message history for the model to maintain reasoning continuity.

Open Source Availability

Model weights are available for download, following Arcee AI's open source approach. The company has published a launch video detailing the model's capabilities and use cases.

What This Means

Trinity Large Thinking enters a competitive reasoning model market currently dominated by OpenAI's o1 and o3 series, Anthropic's extended thinking modes, and DeepSeek's R1. The 262K context window positions it among the larger-context reasoning models available, though it's unclear how this compares to specialized long-context models.

The free tier availability via OpenRouter makes it immediately accessible for developers to test against commercial alternatives. However, without published benchmark scores on standard reasoning evaluations like AIME or GPQA, performance relative to established models remains unverified. The emphasis on "agentic workloads" suggests optimization for multi-step tool use and planning tasks rather than pure mathematical reasoning.

Related Articles

model release

Baidu Launches CoBuddy Code Generation Model with 131K Context Window, Free on OpenRouter

Baidu has released CoBuddy, a code generation model optimized for coding tasks and AI agent workflows. The model features a 131K token context window, up to 65K output tokens, and runs on fp8 quantization with native support for tool calling and reasoning.

model release

IBM Releases Granite Speech 4.1 2B: 2-Billion-Parameter Multilingual Speech Model with Non-Autoregressive Variant

IBM has released Granite Speech 4.1 2B, a 2-billion-parameter speech-language model trained on 174,000 hours of audio for automatic speech recognition and translation across English, French, German, Spanish, Portuguese, and Japanese. The model introduces a dual-head CTC encoder and includes variants for speaker attribution and a novel non-autoregressive architecture for higher throughput.

model release

OpenAI offers EU preview access to GPT-5.5-Cyber model while Anthropic withholds Mythos

OpenAI announced GPT-5.5-Cyber is rolling out in limited preview to vetted cybersecurity teams and is in discussions with the European Commission about preview access. Anthropic released its Mythos model a month ago but has yet to grant EU access for security review.

model release

Supertone releases Supertonic 3: 99M-parameter on-device TTS model supporting 31 languages

Supertone has released Supertonic 3, a 99M-parameter text-to-speech model that runs entirely on-device using ONNX Runtime. The model expands language support from 5 to 31 languages compared to Supertonic 2, requires no GPU, and claims competitive accuracy against models 7-20x larger.

Comments

Loading...