model release

OpenRouter Releases Elephant Alpha: 100B-Parameter Model with 256K Context Window and Free Pricing

TL;DR

OpenRouter has released Elephant Alpha, a 100B-parameter text model with a 256K context window and 32K output token limit. The model is available at no cost through OpenRouter's platform, supporting function calling, structured output, and prompt caching.

2 min read
1

OpenRouter Releases Elephant Alpha: 100B-Parameter Model with 256K Context Window and Free Pricing

OpenRouter has released Elephant Alpha, a 100B-parameter text model designed for "intelligence efficiency" with a 256K context window and support for up to 32K output tokens. The model is available at $0 per million tokens for both input and output through OpenRouter's routing platform.

Technical Specifications

Elephant Alpha features:

  • 100 billion parameters
  • 262,144 (256K) token context window
  • 32,768 (32K) maximum output tokens
  • Function calling support
  • Structured output capabilities
  • Prompt caching
  • Released April 13, 2025

According to OpenRouter, the model focuses on "delivering strong reasoning performance while minimizing token usage," though specific benchmark scores have not been disclosed.

Target Use Cases

OpenRouter positions Elephant Alpha for three primary applications:

  • Code completion and debugging
  • Rapid document processing
  • Lightweight agent interactions

The model is available through OpenRouter's unified API, which routes requests across multiple providers with automatic fallbacks. OpenRouter notes that prompts and completions may be logged by the provider and used for model improvement.

Pricing and Access

The model is currently available at zero cost through OpenRouter's platform, with no charges for input or output tokens. This pricing is managed through OpenRouter's routing system, which normalizes requests and responses across providers.

The model supports OpenAI-compatible API calls and can be accessed through the OpenAI SDK as well as various third-party SDKs and frameworks.

What This Means

Elephant Alpha enters a crowded field of large language models with a distinctive positioning around "intelligence efficiency" and a notably large context window at 256K tokens. The free pricing through OpenRouter makes it accessible for experimentation, though the lack of published benchmarks makes it difficult to assess performance claims against established models. The 32K output token limit is substantially higher than many competing models, which could be useful for document generation tasks. However, the data logging policy and absence of performance metrics warrant careful evaluation for production deployments.

Related Articles

model release

Meta launches Muse Spark model with private API preview and 16 integrated tools

Meta announced Muse Spark today, its first model release since Llama 4 a year ago. The hosted model is available in private API preview and on meta.ai with Instant and Thinking modes, benchmarking competitively against Anthropic's Opus 4.6 and Google's Gemini 3.1 Pro, though behind on Terminal-Bench 2.0.

model release

Anthropic launches Mythos AI model claiming zero-day vulnerability discovery capabilities

Anthropic has launched Mythos, an AI model the company claims can identify and exploit zero-day vulnerabilities with significant capability. The model has not been released publicly, with Anthropic citing security concerns. The announcement raises questions about the model's actual capabilities versus pre-IPO positioning.

model release

Trump officials encourage banks to test Anthropic's Mythos model for security vulnerabilities

U.S. Treasury Secretary Scott Bessent and Federal Reserve Chair Jerome Powell summoned bank executives this week and encouraged them to test Anthropic's newly announced Mythos model for detecting security vulnerabilities. According to Bloomberg, major banks including Goldman Sachs, Citigroup, Bank of America, and Morgan Stanley are already testing the model alongside JPMorgan Chase, despite Anthropic's stated plan to limit initial access.

model release

Arcee AI releases Trinity-Large-Thinking, open reasoning model matching Claude Opus on agent tasks

Arcee AI has released Trinity-Large-Thinking, a 400-billion-parameter open-weight reasoning model with a mixture-of-experts architecture that activates only 13 billion parameters per token. The model matches Claude Opus 4.6 on agent benchmarks like Tau2 and PinchBench but lags on general reasoning tasks. The company spent approximately $20 million—roughly half its total venture capital—to train the model on 2,048 Nvidia B300 GPUs over 33 days.

Comments

Loading...