Facebook iconWhat is Google Gemini CLI & how to install and use it?
Blogs/AI

What is Google Gemini CLI & how to install and use it?

Written by Sharmila Ananthasayanam
Jul 29, 2025
2 Min Read
What is Google Gemini CLI & how to install and use it? Hero

Ever wish your terminal could help you debug, write code, or even run DevOps tasks, without switching tabs? Google’s new Gemini CLI might just do that.

Launched in June 2025, Gemini CLI is an open-source command-line AI tool designed to act like your AI teammate, helping you write, debug, and understand code right from the command line.

What is Gemini CLI?

Gemini CLI is a smart AI assistant you can use directly in your terminal. It’s not just for chatting, it’s purpose-built for developers.

Whether you're reviewing code, fixing bugs, managing Git workflows, or generating docs, Gemini CLI helps you get it done faster. It uses a clever ReAct loop (Reason + Act) to make decisions. This means it doesn’t just give answers, it reasons, acts, and interacts with both your local system and remote model context servers (called MCPs) to perform tasks intelligently.

How to Get Started with Gemini CLI?

Before you begin, make sure you have Node.js v18 or higher installed on your system.

Option 1: Use without installing

You can try it instantly using:

npx https://github.com/google-gemini/gemini-cli

No setup required.

Getting Started with Google Gemini CLI
Learn how to install and use Google Gemini CLI for model interaction, scripting, and automation.
Murtuza Kutub
Murtuza Kutub
Co-Founder, F22 Labs

Walk away with actionable insights on AI adoption.

Limited seats available!

Calendar
Saturday, 14 Mar 2026
10PM IST (60 mins)

Option 2: Install it globally

If you plan to use it often, install it globally:

sudo npm install -g @google/gemini-cli

After installation, just type gemini in your terminal to launch it. 

Installing Gemini globally

Select your desired theme and click Enter.

Authentication Options for Using Gemini CLI

Once Gemini CLI starts, it’ll ask you to authenticate. You’ve got three options:

  1. Login with Google
    • 60 requests per minute
    • 1000 requests per day
  2. Use Gemini API Key
    • Either create a.env file in your project folder and add the line: GEMNI_API_KEY=<your_gemini_api_key> to use your Gemini API key or just enter it in your terminal directly: 

> export GEMINI_API_KEY=<your_gemini_api_key>

  • Free Tier:
    • Flash model only
    • 10 requests/min
    • 250 requests/day
  • Paid Tier:
    • Quota depends on your tier
  1. Use Vertex AI
    • Available for paid accounts
    • Quotas depend on your tier

What Can Gemini CLI Do?

Gemini CLI is more than just an AI chatbot. Here are some of its coolest features:

  • Understand Big CodebasesAsk it questions about your entire codebase, even beyond a 1 million token context window.
  • Generate Apps from PDFs or Sketches Leverage Gemini’s multimodal power to create code from designs or documents.
  • Automate DevOps Tasks Handle pull requests, resolve complex rebases, or run scripts, all with a single prompt.
  • Extend with Tools and MCP Servers Connect tools like Imagen, Veo, or Lyria to add media generation.
  • Ground Queries with Google Search Tap into the world’s knowledge using the built-in Google Search tool.
Getting Started with Google Gemini CLI
Learn how to install and use Google Gemini CLI for model interaction, scripting, and automation.
Murtuza Kutub
Murtuza Kutub
Co-Founder, F22 Labs

Walk away with actionable insights on AI adoption.

Limited seats available!

Calendar
Saturday, 14 Mar 2026
10PM IST (60 mins)

Helpful Tips

  • Press / anytime to open the command helper.
  • Type /docs and hit Enter to open the official Gemini CLI documentation in your browser.

Final Thoughts

Gemini CLI is a smart and flexible tool that helps with coding tasks. It works great for small personal projects or large teamwork, making it easier to write, fix, and manage code right from your terminal.

What makes it stand out?

  • Open-source transparency
  • A generous free tier
  • Advanced AI from Google Gemini
  • Deep integration with developer tools
  • Multimodal support (text, code, images, and more)

And most importantly, it runs in the terminal, your favourite place to work.

Author-Sharmila Ananthasayanam
Sharmila Ananthasayanam

I'm an AIML Engineer passionate about creating AI-driven solutions for complex problems. I focus on deep learning, model optimization, and Agentic Systems to build real-world applications.

Share this article

Phone

Next for you

How Good Is LightOnOCR-2-1B for Document OCR and Parsing? Cover

AI

Mar 6, 202636 min read

How Good Is LightOnOCR-2-1B for Document OCR and Parsing?

Building document processing pipelines is rarely simple. Most OCR systems rely on multiple stages: detection, text extraction, layout parsing, and table reconstruction. When documents become complex, these pipelines often break, making them costly and difficult to maintain. I wanted to understand whether a lightweight end-to-end model could simplify this process without sacrificing document structure. LightOnOCR-2-1B, released by LightOn, takes a different approach. Instead of relying on fragm

How To Build a Voice AI Agent (Using LiveKit)? Cover

AI

Mar 6, 20269 min read

How To Build a Voice AI Agent (Using LiveKit)?

Voice AI agents are becoming increasingly common in applications such as customer support automation, AI call centers, and real-time conversational assistants. Modern voice systems can process speech in real time, understand conversational context, handle interruptions, and respond with natural-sounding speech while maintaining low latency. I wanted to understand what it actually takes to build a production-ready voice AI agent using modern tools. In this guide, I explain how to build a voice

vLLM vs vLLM-Omni: Which One Should You Use? Cover

AI

Mar 10, 20267 min read

vLLM vs vLLM-Omni: Which One Should You Use?

Serving large language models efficiently is a major challenge when building AI applications. As usage scales, systems must handle multiple requests simultaneously while maintaining low latency and high GPU utilization. This is where inference engines like vLLM and vLLM-Omni become important. vLLM is designed to maximize performance for text-based LLM workloads, while vLLM-Omni extends the same architecture to support multimodal inputs such as images, audio, and video. In this guide, we compar