LLxprt Code: Quotas and Pricing
LLxprt Code supports multiple AI providers, each with their own pricing and quota structures. A summary of model usage is available through the /stats command and presented on exit at the end of a session.
Important: Earlier builds surfaced a "Paid Mode" badge in the footer for Gemini sessions. That UI has been removed—LLxprt now stays on the exact model you selected and no longer tries to fall back to flash tiers. To understand whether you're incurring paid traffic, rely on /stats or your provider's billing dashboards.
Provider-Specific Pricing
Google Gemini
For detailed information about Google's quotas and pricing options, including:
- Free tier with Google account login (/auth)
- API key options (paid and unpaid)
- Vertex AI pricing
- Workspace and enterprise options
Please see Google's original documentation for comprehensive details.
Other Providers
Each provider has their own pricing structure:
OpenAI
- Pricing: Pay-per-token model
- Models: o3, o1, GPT-4.1, GPT-4o, and others
- Details: OpenAI Pricing
- Note: All OpenAI usage requires an API key and is paid
Anthropic
- Pricing: Pay-per-token model
- Models: Claude Opus 4, Claude Sonnet 4, and others
- Details: Anthropic Pricing
- Note: All Anthropic usage requires an API key and is paid
Local Models
- Cost: Free (you provide the compute)
- Requirements: Local hardware capable of running the model
- Options: LM Studio, llama.cpp, or any OpenAI-compatible server
OpenRouter
- Pricing: Varies by model (aggregates multiple providers)
- Details: OpenRouter Pricing
- Note: Provides access to 100+ models with unified billing
Fireworks
- Pricing: Competitive rates for fast inference
- Details: Fireworks Pricing
- Note: Optimized for speed and cost-efficiency
Free vs Paid Mode Detection
LLxprt Code attempts to detect when you're in "free" vs "paid" mode:
- Free Mode:
- Google account login with Gemini (limited daily requests)
- Local models
- Paid Mode:
- Any API key usage (except Google's free tier)
- All OpenAI and Anthropic usage
- Most third-party providers
The mode indicator helps you track potential costs, but always verify actual charges with your provider.
Managing Costs
- Monitor Usage: Use
/statsregularly to track token usage - Choose Models Wisely: Smaller models are often cheaper but less capable
- Use Local Models: For development and testing when possible
- Set Budgets: Most providers offer spending limits in their dashboards
Privacy and Terms
See privacy and terms for details on privacy policies and terms of service for each provider.