Choose Your Interface

Nimgir turns fragmented chats into one continuous, peak-quality conversation. No token limits, no degradation.

Desktop

For Claude Desktop users. Structured session continuity with quality awareness built in.

Learn more →

Terminal

For Claude Code users. Configurable quality thresholds and professional workflow tools.

Learn more →

Same engine, different workflows. Pick the one that fits how you work.

All products

Your AI doesn't have memory problems.
It was never designed to remember.

Every new session starts from zero. Your context, your decisions, your reasoning flow — gone. Not because something broke. Because that's how transformers work.

The Facts
Every AI model loses quality as conversations grow longer. This is proven across 18 frontier models. see research →
Your AI doesn't know it's degrading. It delivers confident, worse answers with no warning. see research →
Instructions get forgotten mid-session. The longer the conversation, the more gets skipped. see research →

What Nimgir Does

Nimgir makes every session a continuation of the last one — not just the facts, but the flow.

Continuation

Seamless Continuation

Mid-brainstorm, mid-code, mid-plan — the next session picks up your exact train of thought. Same reasoning style, same direction, same voice. Like talking to the same colleague every day instead of briefing a new one.

Persistence

Decisions That Stay Decided

What you locked stays locked. An append-only log carries your conclusions forward so nothing gets relitigated next session.

Awareness

Degradation-Aware Sessions

Research confirms quality drops as context grows, but nobody publishes the threshold. On Desktop, Nimgir manages sessions around a tested safe zone. On Terminal, you discover your own threshold and lock it in.

Continuity

Cross-Interface

Desktop and Terminal, same workspaces. Brainstorm in the browser, build in the terminal, nothing lost between them.

Already using the free version? The full version doesn't just continue — it remembers why.

Pricing

One-time purchase. No subscription. Yours forever.

Desktop
$17
one-time
For Claude Desktop users. Structured continuation with session-managed quality awareness.

Need more control? See the Terminal version →

Prices are in US Dollars (USD). For other currencies, conversion is handled by your card-issuing bank.

All products

You already know context is finite.
Now you can engineer around it.

Every token competes for the same fixed attention budget. As sessions grow, quality degrades — silently, confidently, invisibly. You've seen it. Nimgir gives you the architecture to manage it.

What the Research Shows
Effective context windows range from 100 to 2,500 tokens across 11 LLMs tested. Up to 99% of the advertised window is unusable for complex reasoning. Paulsen et al., 2025 — MECW see research →
Replacing all irrelevant tokens with whitespace still degraded performance. Input length itself damages reasoning — not content complexity. Du et al., 2025 see research →
When you use AI, you overestimate how well you're performing. Higher AI literacy makes it worse, not better. You have no reliable self-assessment signal. Aalto University, 2026 see research →
Observation masking outperforms LLM summarization for agent context management. Dumping raw context into the window makes things worse, not better. JetBrains / TU Munich, NeurIPS 2025 see research →
Anthropic's own research describes context as a finite resource with diminishing marginal returns — a limited attention budget every token depletes. Anthropic, 2025 see research →
All 18 frontier models tested — GPT-4.1, Claude 4, Gemini 2.5 — grow increasingly unreliable as input length increases. Universal, replicated, architectural. Chroma Research, 2025 see research →

Session Infrastructure, Not a Wrapper

An MCP server that manages context as the finite resource it is. No prompt templates. No magic. Architecture.

Handover

Continuation Briefings

Structured handover documents carry reasoning chains, decision context, and exact stop points across sessions. Not summaries — continuation state.

Persistence

Append-Only Decision Log

Locked conclusions persist across sessions without re-derivation. The log is the source of truth — no drift, no relitigating.

Threshold

Configurable Quality Cap

Set your own token threshold based on your workflow. Sessions split before degradation hits — at a boundary you define and test, not a generic default.

Workspace

Isolated Workspaces

Independent context per project. No cross-contamination. Same workspace accessible from Desktop and Terminal — brainstorm in browser, build in CLI.

Already using the free version? The full version doesn't just continue — it remembers why.

Pricing

One-time purchase. No subscription. Yours forever.

Terminal Team Bundles
7 seats for the price of 5 — save 29%
$495
10 seats for the price of 7 — save 30%
$693
15 seats for the price of 10 — save 33%
$990

Prices are in US Dollars (USD). For other currencies, conversion is handled by your card-issuing bank.

The Research Behind Nimgir

Every claim on this site is backed by published research. No marketing assertions. Here are the sources.

Evidence grades classify each finding's strength: Established means scientific consensus across multiple independent studies. Supported means strong evidence with some debate on scope. Emerging means early evidence that needs further validation.
Emerging Paulsen et al., 2025

Maximum Effective Context Window (MECW)

Across eleven LLMs, the effective context window — the amount of context the model can actually use for complex reasoning — ranges from as low as 100 to 2,500 tokens depending on model and task type. Up to 99% of the advertised context window may be unusable for complex tasks.

Read the research
Established Du et al., 2025

Length Itself Causes Degradation

In a controlled experiment, all non-relevant tokens were replaced with blank spaces — and performance still degraded. This proves that input length itself damages reasoning, not content complexity. The problem is architectural, not solvable by better prompting.

Read the research
Supported Aalto University, Finland

The Reverse Dunning-Kruger Effect

When humans use AI, they consistently overestimate their own performance. Higher AI literacy makes this worse — more technical knowledge brings more confidence, not more accuracy. The traditional Dunning-Kruger effect disappears: everyone overestimates, regardless of skill level. You think you're doing better than you are.

Read the research
Established JetBrains Research & TU Munich — NeurIPS 2025

The Complexity Trap

Observation masking outperforms LLM summarization for agent context management. Raw file contents and exploration threads should be extracted for decisions and status markers, not fed whole into context. More context does not mean better performance — it often means worse.

Read the research
Supported Anthropic, 2025

Context as a Finite Resource

Anthropic's own research describes context as "a finite resource with diminishing marginal returns" — a limited attention budget that every new token depletes. The maker of Claude acknowledges the architectural limitation their product operates under.

Read the research
Established Chroma Research, 2025

Multi-Model Context Degradation Study

Tested 18 frontier models including GPT-4.1, Claude 4, and Gemini 2.5. Every single model's performance grows increasingly unreliable as input length grows. Performance degradation as context length increases is real, universal, and replicated.

Read the research

Download Nimgir

Same installer for demo and full version. Your license key unlocks everything.

Desktop
For Claude Desktop on Windows
v1.0 · .bat
Download Installer

Installs the MCP server locally. No admin rights needed. ~120 KB.

Terminal
For Claude Code
Coming Soon

Also available on GitHub, HuggingFace, and npm.

Setup Guide

Step-by-step walkthrough to get Nimgir running in your Claude environment.

Interactive Setup Wizard

Illustrated guide with copy-paste blocks, platform detection, and license activation. Works for both demo and paid users.

Open Setup Guide →

Documentation

Day-to-day usage guide for Nimgir after installation.

How Nimgir Works

Claude doesn't have memory between conversations. Every new chat starts from zero — your context, decisions, and reasoning chain are gone. This isn't a bug; it's how transformer architecture works.

Nimgir bridges that gap. It's an MCP server that runs locally on your machine and gives Claude structured handover documents at the start of every chat. Instead of starting fresh, Claude picks up exactly where the previous session left off.

The Three Layers

Primer — A fixed set of rules that ship with Nimgir. These tell Claude how to manage sessions: when to wrap up, how to format handover documents, how to track token usage. You never edit the primer — it loads automatically on every get_briefing call.

Briefing — A narrative continuation document written by Claude at the end of each session. It captures the reasoning chain, the exact stop point, open threads, and what the next session should do first. Think of it as a story of how the thinking evolved, not a list of bullet points.

Master Log — An append-only record of locked decisions. Only finalized conclusions go here, tagged with persistent markers that survive across archive cycles. The master log is the single source of truth for what was decided.

Key principle: The briefing carries the narrative thread (what you're thinking about). The master log carries the locked decisions (what you've concluded). Together, they give the next Claude everything it needs to continue without re-explaining.

What Is a Workspace?

A workspace is an isolated container for one project or topic. Each workspace has its own briefing, master log, persistent tags, and archive history. Nothing leaks between workspaces — your web-app project and your research notes never cross-contaminate.

Naming Rules

Workspace names must be lowercase letters, digits, and underscores only, between 1 and 64 characters. Examples: my_project, research_2026, client_alpha.

Creating a Workspace

Workspaces are created automatically the first time you call save_briefing with a new workspace name. If the name isn't registered yet, Nimgir asks Claude to confirm with you before creating it — this catches typos before they become orphan folders.

Listing Workspaces

Ask Claude to call list_workspaces at any time. It returns all registered workspace names with their last-updated dates. Useful for remembering what you named things or checking when a project was last active.

Multiple Workspaces

You can have as many workspaces as you need. To switch between them, just start a new chat and use a different workspace name in your trigger text. Each project or topic gets its own clean context.

The Session Lifecycle

Every Nimgir session follows the same four-step cycle: Load → Work → Wrap → Continue.

1. Load

At the start of each chat, Claude calls get_briefing with your workspace name. Nimgir returns the primer (session rules), the briefing (narrative continuation from last time), and the master log (locked decisions). Claude reads all three and picks up from where you left off — no re-explaining needed.

2. Work

Normal conversation. Claude tracks token usage with a bar at the bottom of each response, showing how much of the quality budget has been used. You work on your project, make decisions, explore ideas — business as usual.

3. Wrap

When the token bar approaches the quality limit (around 55K tokens), Claude lets you know. By 58K, it begins wrapping: writing a new briefing that captures the reasoning chain, locking any new decisions into the master log, and preparing the handover for the next session.

Claude calls save_briefing with the updated briefing and master log. Nimgir writes both files to disk with automatic backups.

4. Continue

Start a new chat. Claude calls get_briefing again and reads the briefing you just saved. The cycle repeats. From your perspective, it's one continuous conversation that never degrades — each segment runs at peak quality.

Why split sessions? Research shows all AI models lose quality as conversations get longer. Nimgir doesn't fight this — it works with it. Short, high-quality segments with structured handovers beat one long degrading session every time.

The Token Bar

Claude posts a token bar after every response, showing estimated usage against the quality budget. The bar starts at 35K (the base context load — primer, briefing, master log, and system prompt consume tokens before any conversation begins).

~42K/60K ████░░░░░░

The Zones

35K – 54K: Safe. Full quality. Work normally.

55K – 57K: Caution. Claude warns you the safe zone is ending and offers to begin wrapping. You can push a bit further if needed, but quality starts thinning.

58K – 60K: Danger. Claude begins the wrap protocol regardless. Past this point, even the handover itself starts losing quality — better to wrap now than risk a degraded continuation.

How Tokens Are Estimated

The rough rule is 4 characters ≈ 1 token. This includes everything in the context window: the system prompt, primer, briefing, master log, your messages, and Claude's responses. The estimation isn't exact, but it's reliable enough to keep sessions in the safe zone.

Tip: If your master log is large (lots of prior decisions), your safe working window shrinks because more tokens are consumed before you start. Archiving old decisions reclaims that space.

"Nimgir tools not found"

Claude can't see the Nimgir MCP server. Check that your Claude Desktop configuration includes the Nimgir server entry, and that the Python path and server script path are correct. Restart Claude Desktop after any config changes.

"Unknown workspace" Warning

You used a workspace name that isn't registered. This is usually a typo. Ask Claude to call list_workspaces to see what names exist. If you intended a new workspace, Claude will ask you to confirm the creation.

Quality Seems to Be Dropping

Check the token bar. If it's above 55K, you're past the safe zone. Start a new chat — the handover will carry your context forward at full quality.

Contact Support

For issues not covered here: info@nimgir.ai

Advanced Guide

Archiving, persistent tags, file locations, data recovery, and license management — included with your license key.

Order Summary

Product
Delivery Digital license key via email
Email
Payment Credit / Debit Card (3D Secure)
Total
← Back to product

Complete Your Purchase

v1.7