Auto-Learn, Context Picker & Seamless Reconnection
Sources in your Library can now learn automatically after indexing, and you can attach existing knowledge — sources, memories, or past conversations — when composing in the Feed. Claude Desktop now works from one bundle across macOS and Windows, remote clients recover gracefully from server outages, update checks respect your platform, and macOS gains a menu-bar background mode.
- Turn on Auto-learn in Settings → Memory Processing, and new sources extract knowledge automatically after indexing — no more clicking Learn one by one
- Attach sources, memories, or past conversations when composing in the Feed — search across all your knowledge, pick what's relevant, and ask questions grounded in context
- Claude Desktop extension now installs on macOS and Windows with no separate Python or Node setup
- Mobile and web apps detect when your server goes offline and show a clear status screen instead of a technical error message
- On macOS, you can now keep Mem in the menu bar after closing the main window; the Dock icon returns automatically when you open the app again
- Labels now propagate to crystals, merged knowledge, and search results — label-based filtering works reliably everywhere
- Auto-learn sources: an opt-in toggle in Settings → Memory Processing that automatically extracts knowledge from new sources after indexing. Manual Learn always takes priority over auto-learn tasks
- Context picker in the Feed: a new button in the Feed toolbar lets you search and attach Library sources, memories, or past conversations when composing. Selected items appear as chips and the resulting memory links back to them
- Labels on ingestion: the upload API, file-path ingestion, and batch endpoints now accept labels — tag sources at the point of capture from the CLI, integrations, or the API
- Claude Desktop extension now installs on macOS and Windows with no separate Python or Node setup. For remote access, it reads the same shared config file as the `nmem` CLI
- New macOS Menu Bar & Dock preference: after you close the main window, Mem can stay in the top menu bar instead of remaining in the Dock. When you reopen the window, the Dock icon comes back automatically
- The `nmem` CLI is now automatically configured when the desktop app generates an API key — no more manually copying credentials
- Connect your phone in seconds: the desktop app can now show a QR code in Access Anywhere settings. Scan it from the mobile or web login screen to connect instantly — no more typing long URLs and API keys
- `nmem status` and the TUI now show whether you're connected locally or to a remote server — no more guessing which Mem you're talking to
- Sources in the Library can now be organized with labels — the same label system used for memories now extends to sources, so you can tag, filter, and group your reference material
- Labels now carry through to crystals and merged knowledge — when Background Intelligence distills or merges memories, the labels from source memories come along. Label-based filtering is now reliable across all knowledge types
- Memory search results now include labels — integrations like Claude Code and Cursor can see and filter by labels in search results
- Mobile and web clients no longer show a raw "Load failed" error when the server is unreachable — a clear reconnection screen appears instead, with automatic retry every few seconds
- If the server goes down while you're using the app, you're taken to the reconnection screen rather than seeing broken views — the app resumes right where you left off once the server is back
- Desktop update checks are now platform-aware. If a new build is available for macOS first, Windows and Linux devices keep seeing their own newest available build instead of a version they cannot install yet
- Thread source filters now work across your whole conversation history. The source menu lists every imported source, and selecting one refreshes the full paginated thread list instead of filtering only the current page
- The Threads source menu now scrolls cleanly when you have many sources, and icons like Gemini CLI, OpenClaw, Antigravity, Grok, Manus, and OpenWebUI render correctly
- Custom providers (SiliconFlow, together.ai, and other OpenAI-compatible services) now work reliably in AI Now — previously they could pass the connection test in Settings but fail when you started a conversation
- AI Now no longer hangs on startup when the agent process fails to launch — the error now appears immediately with diagnostic details instead of an infinite loading spinner
- GitHub Copilot with Codex models (like gpt-5.1-codex-mini) now works in AI Now — these models are automatically routed to the correct API endpoint
- GitHub Copilot headers updated across all integrations to match the current API, fixing authentication errors that appeared with newer Copilot releases
- The desktop app no longer gets stuck on a white screen during startup when the loading view hands off to the main app
- The `nmem` CLI no longer shows a cryptic "Resource doesn't exist" error when the server is unreachable — it now explains the problem clearly and, for first-time remote users, shows exactly how to set up `~/.nowledge-mem/config.json`
- Server URLs with a trailing slash (e.g. `https://mem.example.com/`) no longer cause silent connection failures across the CLI, TUI, and all integrations
