v4.5 preview
Leemer family update

Leemer v4.5 is the detailed drop: research, email, and automation lined up with the exact brains you asked for.

Auto Research and Email Research are the same pipeline until the last mile. You branch and iterate in chat, then flip the send-to-inbox switch when you are ready for a formal memo.

What brief are you brave enough to finish before the sun comes up?

Leemer-only model handoffs

No gradients, just the handoff map we run on every research sprint.

OpenAI

GPT-OSS-120B

Prime

Triages the vibe of every prompt, tags web or file needs, and preps the branch canvas in under a second.

120B activations

OpenRouter

Leemer Deep Research 80B

Investigate

Qwen3 Next 80B Thinking chews through uploads, diagrams, and citations for the deep reasoning lap.

131K context

OpenRouter

Leemer E-Research Pro

Synthesize

Groq-powered scouts fetch sources while Qwen3 Next stitches the email-ready draft and structured citations.

Plan → Cite

Model highlights

MiniMax-M2.1, the new Qwen thinking family, and Kimi K2 Thinking now headline our Leemer lineup.

Fresh lineup

MiniMax-M2.1

Open-source speed

MiniMax-M2.1 keeps our agentic stack lean and lightning-fast without sacrificing composite intelligence. It ranks at the top of the open-source charts for math, coding, and reasoning workloads.

10B activated params · 204K token context

Kimi K2 Thinking

World’s best open-source

MoonshotAI’s Kimi K2 Thinking now powers the toughest planning and multi-step reasoning. We tune it to feel like a GPT-5.1 Chat rival, especially on structured debates and multi-query threads.

MoE + 256K context · beats GPT-5.1 Chat in some benchmarks

Qwen3 235B A22B Thinking

Vision & reasoning

The latest Qwen release keeps our vision-language bench stacked, adding stronger spatial reasoning and diagram-aware dives alongside our other showcase models.

Multi-modal · newly added to the Leemer tier

Auto Research + Email Research = One runway

There is no separate product. Auto Research is where you live, Email Research is what happens when you flip the “Send as Email” switch inside the exact same chat after the findings are ready.

Auto Research work mode

Keep everything inside the same canvas: branch chats, drop screenshots, upload decks, and swap GPT-OSS, Claude, Grok, or Qwen without losing the thread.

  • Live Firecrawl tabs feed evidence straight into the chat you are already in.
  • Keyboard shortcuts + CMD+K jumps to any model or saved brief inside that same space.
  • Share a link without accounts so reviewers can branch it further before you mail it.

Email Research send mode

When the branch is locked, toggle “Mail it” in the same session and we package the exact content into a structured email with citations, attachments, and teammate routing.

  • Pick recipients, CC legal or ops, and reuse your signature block without leaving the chat.
  • Attach PDFs the moment they’re generated with zero downloads or context loss.
  • Track delivery plus replies inside Usage Insights so the same thread shows every status.

Leemer E-Research Pro pipeline

Pulled straight from /research for clarity:

  1. 1.Planning stage identifies every research angle before a single crawl.
  2. 2.Question generator explodes the brief into bite-sized missions.
  3. 3.Parallel research spins up multiple Groq/Perplexity scouts simultaneously.
  4. 4.Final synthesis funnels everything back through Qwen3 Next for citations and email-ready structure.

Token velocity timeline

2023 Launch

9M tokens sparked

9.0M

tokens

YoY changeN/A

2023 Close

40M tokens by winter

40.0M

tokens

YoY change+344.4%

2025 YTD

1B+ tokens already

1.0B

tokens

YoY change+2400.0%

Nine million experimental tokens in early 2023 gave us the data we needed. Forty million tokens by the end of 2023 hardened the pipeline. Surpassing a billion tokens in 2025 is proof that Auto, Email, Writer, and Agents now run on real customer throughput.

Research rituals

  • Branch your chats to compare GPT-OSS-120B with Claude while Firecrawl feeds live web pages.
  • Upload PDFs or decks, then switch to Qwen3 Vision with one click for spatial reasoning.
  • Share any chat without accounts and let Leemer auto-tag the right teammate.
Read the full playbook

Auto Research 2.5

Kicks off with GPT-OSS-120B, pivots to Claude or Grok for debates, and finishes with Leemer email-grade drafts without leaving the workspace.

  • Instant model routing
  • Branch comparisons
  • Built-in email briefs

Deep Focus Stack

Leemer Deep Research 80B and Kimi K2 Thinking team up for multistep plans, diagrams, and document uploads with 256K+ context.

  • Vision reasoning
  • Firecrawl web access
  • Document synthesis

Creator Shortcuts

Keyboard combos, live HTML sandboxes, and drag-in files keep builders shipping prototypes before breakfast.

  • CMD+K everywhere
  • Live Code It sandboxes
  • Upload and cite instantly

Integrations & API

Bring Groq, Zapier, and custom keys. Auto deploys tokens where it counts while Usage Insights keeps the runway clear.

  • Groq plug-in
  • 6k+ Zapier apps
  • Usage safeguards

Themes + Presence

Shadcn inspired theming, installable desktop, and welcome flows that make every new teammate production ready.

  • Theme selector
  • Installable app
  • Custom welcome screen

Email Research Pro

Pick a teammate thread, attach sources, and let the Leemer mailer send curated updates with citations and attachments.

  • Structured citations
  • Recipient targeting
  • Automated follow up

Leemer-only model lineup

Partner models stay inside the selector below. Here’s the Leemer-native lineup with the Lobe icon treatment you asked for.

OpenRouter

Auto Research

Routes GPT-OSS + Claude + Grok

One command fans out to the best brain, keeps Firecrawl feeds alive, and snapshots branches before you swap models mid-thread.

Router
Branch memory
Live web
OpenRouter

Leemer Deep Research 80B

Qwen3 Next 80B Thinking

Multi-hop reasoning for document packs, investor memos, and patent dives. Loves 256K uploads and diagram-heavy briefs.

Long context
Vision aware
Doc ingestion
OpenRouter

Leemer E-Research

Groq compound search

Fast reconnaissance engine that visits pages, extracts tables, and preps structured summaries for Auto or Email Research.

Web agent
Source citations
Firecrawl
OpenRouter

Leemer E-Research Pro

Plan → Qwen3 Next 80B

Planning stage, question generator, parallel Groq runs, then Qwen3 Next synthesis. The exact pipeline lifted from the Research surface.

Planner
Parallel search
Deep synthesis
OpenRouter

Agent-Leemer-K2

Zapier MCP on Groq

Automation agent that can open tickets, update Notion, and mail the recap. Bring your MCP server URL and watch it coordinate 8k+ apps.

Automation
Zapier
Audit trail
OpenRouter

Leemer Heavy

256K dense context

When you need to ingest wikis or legal binders, Heavy keeps the thread stable for hours with steady tone control.

Knowledge base
Batch mode
Tone lock

Cognitive telemetry

Live glyphs stream straight from the same counters that power the token timeline, so the shader sphere is a literal look at Auto Research staying in motion.

2023 -> 9M TOKENS LOGGEDLIVE
2024 -> 40M TOKENS ROUTEDLIVE
2025 -> 1B TOKENS LIVELIVE
AUTO ROUTER: GPT-OSS / CLAUDE / GROKLIVE

Orbit speed

0.75c

Shader speed locked to research refresh cadence.

Signal fidelity

99.2%

Letters and numerals mirror the active branch stack.

Model showcase live embed

Tune the exact model stack that fuels Auto Research, Email Research, the writer, and leaderboard challenges. We kept the designer-friendly chrome, now with Lobe icons on every row.

Curated model lineup

The fastest way to pick the right brain

Compare capabilities, context windows, and sweet spots at a glance. Search anything—from providers to features—and we'll surface the model that fits the job.

Quick search

Showing all 48 models

Multimodal-ready

Image generation limits

Free: 10 images/day, 300/month on `black-forest-labs/flux.2-pro`. Paid: unlimited images plus `black-forest-labs/flux.2-flex`, `openai/gpt-5-image-mini`, and `google/gemini-2.5-flash-image`.

Leemer Labs logo
Flagship agent

Agent‑Leemer‑K2

Orchestrate automations across 8,000+ apps via Zapier MCP. K2 plans, executes, and audits every step so Gmail, Slack, Google Drive, Notion, Asana, and more stay perfectly in sync.

Paste your Zapier MCP Server URL in Settings → API Keys (or the welcome screen) and K2 discovers available tools, enforces permissions, and remembers preferred playbooks.

Zapier MCP
Tool calling
251K ctx
251k tokens context window
Why teams choose K2

Multi-app orchestration

Chain Gmail, Slack, Drive, Notion, and hundreds more in a single mission with typed variables and branching logic.

Zapier MCP automation

Securely execute Zapier MCP actions with scoped permissions, live status tracking, and automatic retries when steps fail.

Human handoff

Generate summaries, approvals, and audit trails so teammates can step in mid-run without losing context.

Guardrails built in

Granular logging, environment pinning, and rate protections keep high-impact automations safe and reviewable.

Setup checklist
  • Visit mcp.zapier.com and sign in (Zapier account required)
  • Create a server and add the tools you want K2 to control
  • Copy the generated Server URL
  • Paste the URL in Settings → API Keys (or on the welcome screen)
Operational tip

Scope access to just the Zapier tools required for each workflow so reviews stay fast and every action log remains crisp.

Leemer Labs logo
Smart default

Auto (Fast)

Automatically selects the best model for your prompt from a curated set of high-quality, fast models. Powered by OpenRouter's Auto Router, it analyzes your prompt and routes to the optimal specialist—from blazing-fast general models to specialized coding and vision models.

Perfect for everyday use when you want smart routing without the overhead of manual model selection. Just say what you need and Auto (Fast) handles the rest.

openrouter/auto-fast
Powered by OpenRouter
Vision + search aware
2.0M tokens context window
Why Auto feels like magic

Instant triage brain

Auto scans the vibe of your prompt—coding bug, research sprint, or quick explainer—and snaps to the right specialist in under a second.

Vision-aware routing

Drop in screenshots or photos and Auto pivots to our vision-language lineup so diagrams, charts, and UI mocks get the attention they deserve.

Fast, deep, or both

Need raw speed, chain-of-thought reasoning, or long-form analysis? Auto balances the trade-offs automatically—no dropdown spelunking required.

One-click default

Set Auto as your daily driver and forget the menu. It front-loads great picks so you can stay in flow while still swapping models on demand.

Daily driver tip

Set Auto as your default model in the chat sidebar. It reads every prompt, routes the moment your cursor leaves the box, and you can still override to a specific model any time.

Vision-first

Stream screenshots, diagrams, or UI mocks and get fast structured breakdowns.

Document-native

Drop in PDFs and research decks—our lineup ingests long-form context in one go.

Agent-friendly

Deterministic reasoning models keep agents reliable, verifiable, and auditable.

Capability legend

VisionPDFFilesSearchReasoningImageThinkingAudioVideo
Partner Models (+ Exclusive)

Our hand-picked lineup for production work: blazing-fast inference, deep context windows, and multi-capability support out of the box.

Qwen

Qwen Plus 0728

Qwen

1.0M tokens
ReasoningThinkingPDFFiles
qwen/qwen-plus-2025-07-28:thinkingClick for details
Qwen

Qwen3 32B

Qwen

131k tokens
ReasoningThinkingPDFFiles
qwen/qwen3-32b:nitroClick for details
Qwen

Qwen3 Max

Qwen

256k tokens
VisionPDFFilesReasoningThinking
qwen/qwen3-maxClick for details
Qwen

Qwen3 VL 235B

Qwen

131k tokens
VisionReasoningPDFFilesThinking
qwen/qwen3-vl-235b-a22b-thinking:nitroClick for details
Qwen

Qwen3 VL 30B

Qwen

131k tokens
VisionPDFFiles
qwen/qwen3-vl-30b-a3b-instruct:nitroClick for details
Google

Gemini 3 Pro

Google

1.0M tokens
VisionAudioVideoPDFFilesReasoning
google/gemini-3-pro-previewClick for details
Google

Gemini 3 Flash

Google

1.0M tokens
VisionAudioVideoPDFFilesReasoningThinking
google/gemini-3-flash-previewClick for details
Mistral

Ministral 3 14B

Mistral

262k tokens
PDFFilesReasoningThinking
mistralai/ministral-14b-2512Click for details
Mistral

Mistral Large 3 2512

Mistral

256k tokens
PDFFilesReasoningThinking
mistralai/mistral-large-2512Click for details
Mistral

Mistral Small Creative

Mistral

131k tokens
ReasoningThinking
mistralai/mistral-small-creativeClick for details
OpenAI

GPT-5 Image Mini

OpenAI

128k tokens
Image
openai/gpt-5-image-miniClick for details
OpenAI

GPT-5 Mini

OpenAI

300k tokens
VisionPDFFilesReasoningThinking
openai/gpt-5-miniClick for details
OpenAI

GPT-5 Nano

OpenAI

128k tokens
VisionPDFFilesReasoning
openai/gpt-5-nanoClick for details
OpenAI

GPT-5.1 Chat

OpenAI

300k tokens
VisionPDFFilesReasoningThinking
openai/gpt-5.1-chatClick for details
OpenAI

GPT-5.2 Chat

OpenAI

400k tokens
VisionPDFFilesReasoningThinking
openai/gpt-5.2-chatClick for details
OpenAI

GPT-OSS-120B

OpenAI

128k tokens
PDFFilesReasoningThinking
openai/gpt-oss-120b:nitroClick for details
OpenAI

GPT-OSS-20B

OpenAI

64k tokens
PDFFilesSearch
openai/gpt-oss-20b:nitroClick for details

FLUX.2 Flex

Black Forest Labs

8k tokens
Image
black-forest-labs/flux.2-flexClick for details

FLUX.2 Pro

Black Forest Labs

8k tokens
Image
black-forest-labs/flux.2-proClick for details
Minimax

MiniMax M2.1

MiniMax

204k tokens
ReasoningThinkingPDFFiles
minimax/minimax-m2.1Click for details
ByteDance

Seed 1.6

ByteDance

256k tokens
VisionPDFFilesReasoningThinking
bytedance-seed/seed-1.6Click for details
ByteDance

Seed 1.6 Flash

ByteDance

256k tokens
VisionPDFFilesReasoningThinking
bytedance-seed/seed-1.6-flashClick for details
Z.ai

GLM-4.6V Nitro

Z.AI

128k tokens
VisionPDFFilesReasoningThinking
z-ai/glm-4.6v:nitroClick for details
Z.ai

GLM-4.7

Z.AI

200k tokens
VisionPDFFilesReasoningThinking
z-ai/glm-4.7Click for details
Kimi

Kimi K2

Moonshot

251k tokens
PDFFiles
moonshotai/kimi-k2-0905:nitroClick for details
Kimi

Kimi K2 Thinking

Moonshot

250k tokens
ReasoningThinkingPDFFiles
moonshotai/kimi-k2-thinking:nitroClick for details
Grok

Grok 4

xAI

200k tokens
VisionPDFFilesReasoningThinking
x-ai/grok-4Click for details
Grok

Grok Code Fast 1

xAI

256k tokens
ReasoningThinkingPDFFiles
x-ai/grok-code-fast-1Click for details
Grok

Grok-4.1-Fast

xAI

2.0M tokens
PDFFilesReasoningThinkingVision
x-ai/grok-4.1-fastClick for details
Anthropic

Claude Haiku 4.5

Anthropic

200k tokens
VisionPDFFilesReasoningThinking
anthropic/claude-haiku-4.5Click for details
Anthropic

Claude Sonnet 4.5

Anthropic

200k tokens
VisionPDFFilesReasoningThinking
anthropic/claude-sonnet-4.5Click for details
DeepSeek

DeepSeek V3.2 Speciale

DeepSeek

131k tokens
ReasoningThinkingFiles
deepseek/deepseek-v3.2:nitroClick for details
Stepfun

Step3

Stepfun

66k tokens
VisionReasoningPDFFiles
stepfun-ai/step3Click for details
AWS

Amazon Nova 2 Lite

Amazon

1.0M tokens
VisionVideoPDFFilesReasoning
amazon/nova-2-lite-v1Click for details
AWS

Amazon Nova Premier

Amazon

1.0M tokens
VisionPDFFiles
amazon/nova-premier-v1Click for details
Xiaomi

MiMo-V2-Flash

Xiaomi

256k tokens
ReasoningThinking
xiaomi/mimo-v2-flash:freeClick for details
Azure

TNG R1T Chimera

Microsoft

200k tokens
ReasoningThinkingPDFFiles
tngtech/tng-r1t-chimeraClick for details
Nvidia

Nemotron 3 Nano 30B

NVIDIA

131k tokens
ReasoningThinking
nvidia/nemotron-3-nano-30b-a3bClick for details
Nvidia

Nemotron Nano 2 VL

NVIDIA

131k tokens
VisionPDFFilesThinking
nvidia/nemotron-nano-12b-v2-vlClick for details
OpenRouter

Auto (Fast)

OpenRouter

2.0M tokens
VisionSearchPDFFilesReasoning
openrouter/auto-fastClick for details
OpenRouter

Auto (MAX)

OpenRouter

2.0M tokens
VisionSearchPDFFilesReasoningThinking
openrouter/auto-maxClick for details
OpenRouter

Auto (Research)

OpenRouter

2.0M tokens
VisionSearchPDFFilesReasoning
openrouter/auto-researchClick for details
Google Gemini

Long-context multimodal models with native Google search grounding and file understanding.

Google

Gemini 2.5 Flash Image

Google

1.0M tokens
Image
gemini-2.5-flash-image-previewClick for details
Google

Gemini 2.5 Flash

Google

1.0M tokens
VisionSearchPDFFiles
gemini-2.5-flashClick for details
Google

Gemini 2.5 Pro

Google

2.1M tokens
VisionSearchPDFFilesReasoning
gemini-2.5-proClick for details
Groq Speed Lab

Low-latency models riding Groq's tensor streaming engine—perfect when every millisecond counts.

Groq

DeepSeek R1 Distill Llama 70B

Groq

128k tokens
ReasoningThinking
deepseek-r1-distill-llama-70bClick for details
Groq

Llama 4 Maverick

Groq

131k tokens
General-purpose text generation
meta-llama/llama-4-maverick-17b-128e-instructClick for details
Groq

Qwen 3 32B

Groq

33k tokens
ReasoningThinking
qwen/qwen3-32bClick for details

Letter from the CEO

Crew, v4.5 is a love letter to every person who DM'd us screenshots, bug repros, and wild feature threads. You asked for clearer boundaries between Auto and Email Research so we opened the pipeline, you wanted our own models front and center so we shipped the Leemer-only lineup, and you kept pushing for better visuals so the new brain sphere is literally powered by shader math.

We are working ridiculous hours to make sure GPT-OSS-120B, Leemer Deep Research 80B, Groq scouts, and Qwen3 Next behave like a single organism. Next up is the Leemer API preview so you can hit one endpoint and get mixture-of-model routing in a single request. It is the same orchestration we run internally, now packaged for your own stacks.

We also moved everyone onto the Autumn pricing track so bills stay predictable while we keep sponsoring open source research, docs, and demo repos that inspire us. When you see a new OSS project carrying the Leemer badge, it is because your subscriptions made it happen.

Need anything, want to test the API preview, or have an idea we missed? Email dev@leemerchat.com and the core team (including me) will answer.

Stay relentless, stay curious, and ship the message no one expects.

Repath Khan

Founder and CEO