LLM Providers
Supported remote LLM providers, model guidance, and known caveats for AI-Now and agent workflows
Choose a provider based on your workflow, then configure it once in Settings (or nmem CLI/TUI).
Nowledge Mem recommends subscription-based providers for the smoothest daily experience:
- OpenAI ChatGPT/Codex subscription
- Kimi Code subscription
Cost-aware default recommendation
For day-to-day AI-Now usage, prioritize fast tool-capable models over SOTA by default.
Examples: gpt-5.1-codex-mini (Codex subscription) or Kimi Coding Plan models.
If your current setup is spending too many tokens (for example always running gpt-5.3-codex), switch to a lighter tool-capable default first.
What To Choose
ChatGPT/Codex Subscription
Best default for coding-heavy AI-Now workflows with strong tool use support.
Kimi Code Subscription
Great for coding workflows and tool-calling in AI-Now.
Provider Guides
- Stable tool-calling in AI-Now and agents
- Strong ecosystem compatibility
- Great fit for coding-heavy AI-Now sessions
- Reliable tool workflows in day-to-day usage
- Strong planning and tool use quality
- Good for long multi-step sessions
- Use
deepseek-chatin AI-Now - Solid lower-cost alternative for tool workflows
- Flexible multi-model routing
- Pick tool-capable models for AI-Now/agents
- Supported in AI-Now and agent flows
- Smooth fit with Google AI Studio credentials
- Supported with tool workflows in AI-Now
- Good when xAI is already part of your stack
- Supported in AI-Now and extension workflows
- Works with current MiniMax chat models
- Supported for AI-Now and agents
- Good regional/provider fit for existing usage
- No cloud dependency for model runtime
- Use tool-capable models for AI-Now agents
- Supported in AI-Now
- Low-friction if your team already uses Copilot
- Requires OpenAI-compatible chat completions
- Tool support depends on your gateway/model
DeepSeek model hint
Use deepseek-chat for AI-Now and agent tasks.
Custom Provider Guidance
If you configure a custom OpenAI-compatible endpoint (openai_compatible) that points to DeepSeek (api.deepseek.com), use deepseek-chat as the model for AI-Now and agent tasks.
For Linux headless deployment setup, see Linux Server Deployment.