Skip to content

Model Selection

Choose the right Gemini model for your task.

Available Models

Gemini-2.5-pro

  • Best for: Complex analysis, large codebases
  • Context: 2M tokens
  • Use when: Analyzing entire projects, architectural reviews, stronger reasoning

Gemini-2.5-flash

  • Best for: Quick responses, routine tasks
  • Context: 1M tokens
  • Use when: Fast code reviews, Analyzing entire projects, simple explanations

Setting Models

bash
You need use natural language: "...using gemini flash"
bash
You can also append with '-m' or ask specifically with

In Configuration

json
{
  "mcpServers": {
    "gemini-cli": {
      "command": "gemini-mcp",
      "env": {
        "GEMINI_MODEL": "gemini-1.5-flash"
      }
    }
  }
}

Per Request (Coming Soon)

/gemini-cli:analyze --model=flash @file.js quick review

Model Comparison

ModelSpeedContextBest Use Case
ProSlower2M tokensbig ideas
FlashFast1M tokensquick, specific changes

Cost Optimization

  1. Start with Flash for most tasks
  2. Use Pro only when you need the full context
  3. Flash-8B for simple, repetitive tasks

Token Limits

  • Pro: ~2 million tokens (~500k lines of code)
  • Flash: ~1 million tokens (~250k lines of code)
  • Flash-8B: ~1 million tokens (~250k lines of code)

Recommendations

  • Code Review: Flash
  • Architecture Analysis: Pro
  • Quick Fixes: Flash-8B
  • Documentation: Flash
  • Security Audit: Pro

Released under the MIT License.