Translation Providers
The AI Localization Automator supports five different AI providers, each with unique strengths and configuration options. Choose the provider that best fits your project's needs, budget, and quality requirements.
Ollama (Local AI)
Best for: Privacy-sensitive projects, offline translation, unlimited usage
Ollama runs AI models locally on your machine, providing complete privacy and control without API costs or internet requirements.
Popular Models
- llama3.2 (Recommended general purpose)
- mistral (Efficient alternative)
- codellama (Code-aware translations)
- And many more community models
Configuration Options
- Base URL: Local Ollama server (default:
http://localhost:11434
) - Model: Name of locally installed model (required)
- Use Chat API: Enable for better conversation handling
- Temperature: 0.0-2.0 (0.3 recommended)
- Max Tokens: 1-8,192 tokens
- Context Size: 512-32,768 tokens
- Request Timeout: 10-300 seconds (local models can be slower)
- Enable Streaming: For real-time response processing
Strengths
- ✅ Complete privacy (no data leaves your machine)
- ✅ No API costs or usage limits
- ✅ Works offline
- ✅ Full control over model parameters
- ✅ Wide variety of community models
- ✅ No vendor lock-in
Considerations
- 💻 Requires local setup and capable hardware
- ⚡ Generally slower than cloud providers
- 🔧 More technical setup required
- 📊 Translation quality varies significantly by model (some can exceed cloud providers)
- 💾 Large storage requirements for models
Setting Up Ollama
- Install Ollama: Download from ollama.ai and install on your system
- Download Models: Use
ollama pull llama3.2
to download your chosen model - Start Server: Ollama runs automatically, or start with
ollama serve
- Configure Plugin: Set base URL and model name in the plugin settings
- Test Connection: The plugin will verify connectivity when you apply configuration
OpenAI
Best for: Highest overall translation quality, extensive model selection
OpenAI provides industry-leading language models through their API, including the latest GPT models and the new Responses API format.
Available Models
- gpt-5 (Latest flagship model)
- gpt-5-mini (Smaller, faster variant)
- gpt-4.1 and gpt-4.1-mini
- gpt-4o and gpt-4o-mini (Optimized models)
- o3 and o3-mini (Advanced reasoning)
- o1 and o1-mini (Previous generation)
Configuration Options
- API Key: Your OpenAI API key (required)
- Base URL: API endpoint (default: OpenAI's servers)
- Model: Choose from available GPT models
- Temperature: 0.0-2.0 (0.3 recommended for translation consistency)
- Max Output Tokens: 1-128,000 tokens
- Request Timeout: 5-300 seconds
- Verbosity: Control response detail level
Strengths
- ✅ Consistently high-quality translations
- ✅ Excellent context understanding
- ✅ Strong format preservation
- ✅ Wide language support
- ✅ Reliable API uptime
Considerations
- 💰 Higher cost per request
- 🌐 Requires internet connection
- ⏱️ Usage limits based on tier
Anthropic Claude
Best for: Nuanced translations, creative content, safety-focused applications
Claude models excel at understanding context and nuance, making them ideal for narrative-heavy games and complex localization scenarios.
Available Models
- claude-opus-4-1-20250805 (Latest flagship)
- claude-opus-4-20250514
- claude-sonnet-4-20250514
- claude-3-7-sonnet-20250219
- claude-3-5-haiku-20241022 (Fast and efficient)
- claude-3-haiku-20240307
Configuration Options
- API Key: Your Anthropic API key (required)
- Base URL: Claude API endpoint
- Model: Select from Claude model family
- Temperature: 0.0-1.0 (0.3 recommended)
- Max Tokens: 1-64,000 tokens
- Request Timeout: 5-300 seconds
- Anthropic Version: API version header
Strengths
- ✅ Exceptional context awareness
- ✅ Great for creative/narrative content
- ✅ Strong safety features
- ✅ Detailed reasoning capabilities
- ✅ Excellent instruction following
Considerations
- 💰 Premium pricing model
- 🌐 Internet connection required
- 📏 Token limits vary by model
DeepSeek
Best for: Cost-effective translation, high throughput, budget-conscious projects
DeepSeek offers competitive translation quality at a fraction of the cost of other providers, making it ideal for large-scale localization projects.
Available Models
- deepseek-chat (General purpose, recommended)
- deepseek-reasoner (Enhanced reasoning capabilities)
Configuration Options
- API Key: Your DeepSeek API key (required)
- Base URL: DeepSeek API endpoint
- Model: Choose between chat and reasoner models
- Temperature: 0.0-2.0 (0.3 recommended)
- Max Tokens: 1-8,192 tokens
- Request Timeout: 5-300 seconds
Strengths
- ✅ Very cost-effective
- ✅ Good translation quality
- ✅ Fast response times
- ✅ Simple configuration
- ✅ High rate limits
Considerations
- 📏 Lower token limits
- 🆕 Newer provider (less track record)
- 🌐 Requires internet connection
Google Gemini
Best for: Multilingual projects, cost-effective translation, Google ecosystem integration
Gemini models offer strong multilingual capabilities with competitive pricing and unique features like thinking mode for enhanced reasoning.
Available Models
- gemini-2.5-pro (Latest flagship with thinking)
- gemini-2.5-flash (Fast, with thinking support)
- gemini-2.5-flash-lite (Lightweight variant)
- gemini-2.0-flash and gemini-2.0-flash-lite
- gemini-1.5-pro and gemini-1.5-flash
Configuration Options
- API Key: Your Google AI API key (required)
- Base URL: Gemini API endpoint
- Model: Select from Gemini model family
- Temperature: 0.0-2.0 (0.3 recommended)
- Max Output Tokens: 1-8,192 tokens
- Request Timeout: 5-300 seconds
- Enable Thinking: Activate enhanced reasoning for 2.5 models
- Thinking Budget: Control thinking token allocation
Strengths
- ✅ Strong multilingual support
- ✅ Competitive pricing
- ✅ Advanced reasoning (thinking mode)
- ✅ Google ecosystem integration
- ✅ Regular model updates
Considerations
- 🧠 Thinking mode increases token usage
- 📏 Variable token limits by model
- 🌐 Internet connection required
Choosing the Right Provider
Provider | Best For | Quality | Cost | Setup | Privacy |
---|---|---|---|---|---|
Ollama | Privacy/offline | Variable* | Free | Advanced | Local |
OpenAI | Highest quality | ⭐⭐⭐⭐⭐ | 💰💰💰 | Easy | Cloud |
Claude | Creative content | ⭐⭐⭐⭐⭐ | 💰💰💰💰 | Easy | Cloud |
DeepSeek | Budget projects | ⭐⭐⭐⭐ | 💰 | Easy | Cloud |
Gemini | Multilingual | ⭐⭐⭐⭐ | 💰 | Easy | Cloud |
*Quality for Ollama varies significantly based on the local model used - some modern local models can match or exceed cloud providers.
Provider Configuration Tips
For All Cloud Providers:
- Store API keys securely and don't commit them to version control
- Start with conservative temperature settings (0.3) for consistent translations
- Monitor your API usage and costs
- Test with small batches before large translation runs
For Ollama:
- Ensure adequate RAM (8GB+ recommended for larger models)
- Use SSD storage for better model loading performance
- Consider GPU acceleration for faster inference
- Test locally before relying on it for production translations