انتقل إلى المحتوى الرئيسي

Translation Providers

The AI Localization Automator supports five different AI providers, each with unique strengths and configuration options. Choose the provider that best fits your project's needs, budget, and quality requirements.

Ollama (Local AI)

Best for: Privacy-sensitive projects, offline translation, unlimited usage

Ollama runs AI models locally on your machine, providing complete privacy and control without API costs or internet requirements.

  • llama3.2 (Recommended general purpose)
  • mistral (Efficient alternative)
  • codellama (Code-aware translations)
  • And many more community models

Configuration Options

  • Base URL: Local Ollama server (default: http://localhost:11434)
  • Model: Name of locally installed model (required)
  • Use Chat API: Enable for better conversation handling
  • Temperature: 0.0-2.0 (0.3 recommended)
  • Max Tokens: 1-8,192 tokens
  • Context Size: 512-32,768 tokens
  • Request Timeout: 10-300 seconds (local models can be slower)
  • Enable Streaming: For real-time response processing

Strengths

  • ✅ Complete privacy (no data leaves your machine)
  • ✅ No API costs or usage limits
  • ✅ Works offline
  • ✅ Full control over model parameters
  • ✅ Wide variety of community models
  • ✅ No vendor lock-in

Considerations

  • 💻 Requires local setup and capable hardware
  • ⚡ Generally slower than cloud providers
  • 🔧 More technical setup required
  • 📊 Translation quality varies significantly by model (some can exceed cloud providers)
  • 💾 Large storage requirements for models

Setting Up Ollama

  1. Install Ollama: Download from ollama.ai and install on your system
  2. Download Models: Use ollama pull llama3.2 to download your chosen model
  3. Start Server: Ollama runs automatically, or start with ollama serve
  4. Configure Plugin: Set base URL and model name in the plugin settings
  5. Test Connection: The plugin will verify connectivity when you apply configuration

OpenAI

Best for: Highest overall translation quality, extensive model selection

OpenAI provides industry-leading language models through their API, including the latest GPT models and the new Responses API format.

Available Models

  • gpt-5 (Latest flagship model)
  • gpt-5-mini (Smaller, faster variant)
  • gpt-4.1 and gpt-4.1-mini
  • gpt-4o and gpt-4o-mini (Optimized models)
  • o3 and o3-mini (Advanced reasoning)
  • o1 and o1-mini (Previous generation)

Configuration Options

  • API Key: Your OpenAI API key (required)
  • Base URL: API endpoint (default: OpenAI's servers)
  • Model: Choose from available GPT models
  • Temperature: 0.0-2.0 (0.3 recommended for translation consistency)
  • Max Output Tokens: 1-128,000 tokens
  • Request Timeout: 5-300 seconds
  • Verbosity: Control response detail level

Strengths

  • ✅ Consistently high-quality translations
  • ✅ Excellent context understanding
  • ✅ Strong format preservation
  • ✅ Wide language support
  • ✅ Reliable API uptime

Considerations

  • 💰 Higher cost per request
  • 🌐 Requires internet connection
  • ⏱️ Usage limits based on tier

Anthropic Claude

Best for: Nuanced translations, creative content, safety-focused applications

Claude models excel at understanding context and nuance, making them ideal for narrative-heavy games and complex localization scenarios.

Available Models

  • claude-opus-4-1-20250805 (Latest flagship)
  • claude-opus-4-20250514
  • claude-sonnet-4-20250514
  • claude-3-7-sonnet-20250219
  • claude-3-5-haiku-20241022 (Fast and efficient)
  • claude-3-haiku-20240307

Configuration Options

  • API Key: Your Anthropic API key (required)
  • Base URL: Claude API endpoint
  • Model: Select from Claude model family
  • Temperature: 0.0-1.0 (0.3 recommended)
  • Max Tokens: 1-64,000 tokens
  • Request Timeout: 5-300 seconds
  • Anthropic Version: API version header

Strengths

  • ✅ Exceptional context awareness
  • ✅ Great for creative/narrative content
  • ✅ Strong safety features
  • ✅ Detailed reasoning capabilities
  • ✅ Excellent instruction following

Considerations

  • 💰 Premium pricing model
  • 🌐 Internet connection required
  • 📏 Token limits vary by model

DeepSeek

Best for: Cost-effective translation, high throughput, budget-conscious projects

DeepSeek offers competitive translation quality at a fraction of the cost of other providers, making it ideal for large-scale localization projects.

Available Models

  • deepseek-chat (General purpose, recommended)
  • deepseek-reasoner (Enhanced reasoning capabilities)

Configuration Options

  • API Key: Your DeepSeek API key (required)
  • Base URL: DeepSeek API endpoint
  • Model: Choose between chat and reasoner models
  • Temperature: 0.0-2.0 (0.3 recommended)
  • Max Tokens: 1-8,192 tokens
  • Request Timeout: 5-300 seconds

Strengths

  • ✅ Very cost-effective
  • ✅ Good translation quality
  • ✅ Fast response times
  • ✅ Simple configuration
  • ✅ High rate limits

Considerations

  • 📏 Lower token limits
  • 🆕 Newer provider (less track record)
  • 🌐 Requires internet connection

Google Gemini

Best for: Multilingual projects, cost-effective translation, Google ecosystem integration

Gemini models offer strong multilingual capabilities with competitive pricing and unique features like thinking mode for enhanced reasoning.

Available Models

  • gemini-2.5-pro (Latest flagship with thinking)
  • gemini-2.5-flash (Fast, with thinking support)
  • gemini-2.5-flash-lite (Lightweight variant)
  • gemini-2.0-flash and gemini-2.0-flash-lite
  • gemini-1.5-pro and gemini-1.5-flash

Configuration Options

  • API Key: Your Google AI API key (required)
  • Base URL: Gemini API endpoint
  • Model: Select from Gemini model family
  • Temperature: 0.0-2.0 (0.3 recommended)
  • Max Output Tokens: 1-8,192 tokens
  • Request Timeout: 5-300 seconds
  • Enable Thinking: Activate enhanced reasoning for 2.5 models
  • Thinking Budget: Control thinking token allocation

Strengths

  • ✅ Strong multilingual support
  • ✅ Competitive pricing
  • ✅ Advanced reasoning (thinking mode)
  • ✅ Google ecosystem integration
  • ✅ Regular model updates

Considerations

  • 🧠 Thinking mode increases token usage
  • 📏 Variable token limits by model
  • 🌐 Internet connection required

Choosing the Right Provider

ProviderBest ForQualityCostSetupPrivacy
OllamaPrivacy/offlineVariable*FreeAdvancedLocal
OpenAIHighest quality⭐⭐⭐⭐⭐💰💰💰EasyCloud
ClaudeCreative content⭐⭐⭐⭐⭐💰💰💰💰EasyCloud
DeepSeekBudget projects⭐⭐⭐⭐💰EasyCloud
GeminiMultilingual⭐⭐⭐⭐💰EasyCloud

*Quality for Ollama varies significantly based on the local model used - some modern local models can match or exceed cloud providers.

Provider Configuration Tips

For All Cloud Providers:

  • Store API keys securely and don't commit them to version control
  • Start with conservative temperature settings (0.3) for consistent translations
  • Monitor your API usage and costs
  • Test with small batches before large translation runs

For Ollama:

  • Ensure adequate RAM (8GB+ recommended for larger models)
  • Use SSD storage for better model loading performance
  • Consider GPU acceleration for faster inference
  • Test locally before relying on it for production translations