Poolside Launches Laguna M.1, Free-Tier Coding Agent Model with 128K Context Window
Poolside has released Laguna M.1, its flagship coding agent model available for free on OpenRouter. The model features a 128K context window, up to 8K output tokens, and is optimized for agentic coding workflows with tool calling and reasoning capabilities.
Poolside Launches Laguna M.1, Free-Tier Coding Agent Model with 128K Context Window
Poolside has released Laguna M.1, its flagship coding agent model, available for free through OpenRouter. The model is designed for complex software engineering tasks and agentic coding workflows.
Key Specifications
Laguna M.1 offers a 128K context window (131,072 tokens) and supports up to 8K output tokens. The model has been quantized to fp8 (8-bit floating point) for efficient inference. According to Poolside, the model is optimized for tool calling and reasoning capabilities specifically tailored to software development workflows.
The free tier on OpenRouter charges $0 per million tokens for both input and output, making it accessible for developers to test and integrate into their coding workflows.
Technical Implementation
The model supports reasoning-enabled workflows, allowing it to show step-by-step thinking processes through OpenRouter's reasoning parameter. Developers can access the model's internal reasoning through the reasoning_details array in API responses.
Poolside has positioned Laguna M.1 as an agentic model, meaning it's designed to operate autonomously within coding environments with tool calling capabilities rather than simple code completion.
Release Timeline
Laguna M.1 was released on April 28, 2026, according to OpenRouter's listing. The model is currently available exclusively through OpenRouter's API, which provides normalized requests and responses across multiple providers.
What This Means
Poolside's entry into the coding agent space with a free-tier model signals increasing competition in AI-assisted software development. The 128K context window is substantial for code-related tasks, allowing developers to work with large codebases or complex multi-file operations. The fp8 quantization suggests Poolside is prioritizing inference efficiency, likely to maintain the free tier sustainably. However, without published benchmarks on standard coding evaluations like HumanEval or MBPP, it remains unclear how Laguna M.1's performance compares to established coding models from GitHub Copilot, Amazon CodeWhisperer, or Replit's offerings.
Related Articles
Poolside releases Laguna XS.2, free fp8-quantized coding agent with 128K context
Poolside has released Laguna XS.2, the second-generation model in its XS size class for agentic coding workflows. The model offers 128K context window, up to 8K output tokens, and is quantized to fp8 for efficiency, available free via OpenRouter.
Moonshot AI Launches 'Kimi Latest' Router Model with 262K Context Window
Moonshot AI released Kimi Latest, a router endpoint that automatically redirects to the most recent model in the Kimi family. The model features a 262,144 token context window, though specific pricing and performance benchmarks have not been disclosed.
NVIDIA Nemotron 3 Nano Omni: 30B-parameter multimodal model launches on AWS SageMaker with 131K token context
NVIDIA has launched Nemotron 3 Nano Omni on Amazon SageMaker JumpStart, a multimodal model with 30 billion total parameters (3 billion active) that processes video, audio, images, and text in a single inference pass. The model features a 131K token context window and uses a Mamba2 Transformer Hybrid MoE architecture combining three specialized encoders.
Nvidia releases Nemotron 3 Nano Omni: 30B-parameter multimodal model with 256K context, free on OpenRouter
Nvidia has released Nemotron 3 Nano Omni, a 30-billion-parameter multimodal model available free on OpenRouter. The model features a 256,000-token context window, accepts text, image, video, and audio inputs, and claims 2× higher throughput for video reasoning compared to separate vision and speech pipelines.
Comments
Loading...