Gemini 3.1 Pro API Pricing: $1/M Input, $6/M Output – Low-Cost Grsai Access Guide for Developers

 Have you used Gemini 3.1 Pro yet? Google says its reasoning ability has more than doubled compared to Gemini 3 Pro. Is it really that impressive? This article will comprehensively introduce the core capabilities of Gemini 3.1 Pro, how developers can call the Gemini 3.1 Pro API, and how ordinary users can quickly use it.

Press enter or click to view image in full size

I. Gemini 3.1 Pro Builds an All-Round Reasoning Brain What exactly does

ARC-AGI-2 test, and what does 77% mean?

This benchmark test does not assess knowledge reserves but reasoning ability. What makes this test special is: it only gives you a few grid patterns as ‘examples,’ and you need to deduce the underlying transformation rules yourself, then apply them to entirely new patterns.

For humans, this kind of reasoning can be completed almost intuitively, while AI cannot rely on rote memorization or statistical guessing — it must truly understand the logic. Because it gets stuck in areas where ‘humans excel and AI struggles,’ ARC-AGI has always been regarded as an important measure of progress toward general artificial intelligence (AGI).

The number 77.1% behind it shows that Google has found an effective path to scale reasoning capabilities. While other models are still relying on piling up parameters, Gemini 3.1 Pro is far ahead in ‘understanding logic and deducing rules.’

Press enter or click to view image in full size

How does the three-level thinking mode save money?

Gemini 3.1 Pro adds an intermediate tier, upgrading to a three-level thinking system. Developers can flexibly switch between low, medium, and high via the ‘thinking_level’ parameter. Each tier has an internal ‘thinking Token’ budget, which directly determines how much effort the model spends on internal reasoning. For simple API calls, use the low tier to save latency and cost; for complex debugging, switch to the high tier.

Low: Fast response mode, with the lowest thinking Token budget. The model responds quickly based on intuition, suitable for daily simple tasks. It gives the most direct answers based on intuition and training data, with the shortest thinking process. Compared to the previous generation’s Fast mode, response speed is faster and costs are lower — for simple tasks, it can save 60%-80% of reasoning costs compared to using the high tier.

Medium: Balanced reasoning mode, with moderate thinking Token budget, performing medium-depth logical analysis. The reasoning quality is equivalent to the previous generation model (Gemini 3 Pro)’s ‘high’ tier, but the cost is only about 40% of that. This means you can achieve the same complexity as before with much less money.

High: Deep reasoning mode (Deep Think Mini), with the highest thinking Token budget, suitable for professional complex tasks. The model activates a powerful reasoning engine similar to ‘Deep Think,’ performing extremely in-depth multi-step thinking to achieve the best results. Although the cost is higher than the low and medium tiers, compared to the previous generation’s Deep Think mode, the cost for the same depth is reduced by about 30% — because the architecture is optimized, the money is spent more efficiently.

Press enter or click to view image in full size

Native multimodal support, comprehensive expansion of perception

Seamless cross-modal information fusion: True multimodality means establishing connections between different modalities. Gemini 3.1 Pro can perform cross-reasoning on video frames, audio narration, and text subtitles, building understanding of complex tasks on richer perception dimensions.

Press enter or click to view image in full size

Lower hallucination rate, knowing what it doesn’t know

Significance for professional reasoning: In rigorous scenarios like code debugging and academic research, ‘I don’t know’ is safer than ‘guessing wildly.’ This capability comes from the model’s clear modeling of knowledge boundaries. In HIGH tier, the model spends more time on internal verification, further reducing hallucination risks in critical tasks.

II. Real Performance of Gemini 3.1 Pro:

Code Generation for SVG Animations

Press enter or click to view image in full size

Integrating Complex Systems

Interaction Design

Press enter or click to view image in full size

Creative Programming

Press enter or click to view image in full size

III. How to Use Gemini 3.1 Pro, Low-Cost Official API

Press enter or click to view image in full size

Gemini Official Source API — Grsai Although the API pricing for Gemini 3.1 Pro remains the same as the previous generation 3.0 Pro (input $2/M tokens, output $12/M tokens), this price is still not low for individual users and enterprises. Third-party APIs on the market often have inflated prices with middlemen taking cuts. In fact, you can directly call Gemini’s source API — GrsAi (https://grsai.com, which allows calling Gemini-3.1-Pro, Nano Banana Pro (image generation), and other models at lower prices than official.


Image generation models:

  • Nano Banana 2 — $0.065/image
  • Nano Banana Pro — $0.09/image
  • Gpt-image-1.5 — $0.02/image
  • Nano Banana — $0.022/image

Chat models:

  • Gemini-3.1-Pro: input $1~$2 /M tokens, output $6~$12 /M tokens
  • Gemini-3-Pro: input $1~$2 /M tokens, output $6~$12 /M tokens
  • Gemini-2.5-Pro: input $1.25~$2.25 /M tokens, output $6.25~$12.5 /M tokens

Video models:

  • Sora2 — $0.08/clip
  • Veo3.1 — $0.4/clip

For more model versions, check the model list: https://grsai.com/dashboard/models

IV. Development Access to Gemini 3.1 Pro API Tutorial API Node

  • Overseas: https://grsaiapi.com
  • Domestic direct connection: https://grsai.dakka.com.cn
  • Host+ interface: https://grsai.dakka.com.cn/v1/draw/nano-banana
Press enter or click to view image in full size

Chat API Call Instructions

Method 1: Call via Grsai (OpenAI compatible format)

import openai
# Configure Grsai address and key
client = openai.OpenAI(
api_key="your APIKey",
base_url="https://grsai.dakka.com.cn/v1" # Domestic direct connection address
)
# Call Gemini-3.1-Pro
response = client.chat.completions.create(
model="gemini-3.1-pro",
messages=[
{"role": "system", "content": "You are an aerospace data expert"},
{"role": "user", "content": "Analyze the International Space Station's orbital data"}
],
stream=False
)
print(response.choices[0].message.content)

Method 2: Call via Google official SDK

  • Replaced interface address: https://grsai.dakka.com.cn/v1beta/models/gemini-3.1-pro:generateContent (model name must match Grsai backend)
  • Gemini API official documentation: https://ai.google.dev/gemini-api/docs
import requests
import json
# Grsai domestic direct connection node + Gemini native interface path
url = "https://grsai.dakka.com.cn/v1beta/models/gemini-3.1-pro:generateContent"
# Your Grsai API Key
api_key = "your Grsai-APIKey"
# Request headers
headers = {
"Content-Type": "application/json"
}
# Request body (Gemini native format)
data = {
"contents": [
{
"parts": [
{
"text": "Explain what a vector graphic is"
}
]
}
]
}
# Send request (API Key passed via URL parameter, Gemini style)
response = requests.post(
f"{url}?key={api_key}",
headers=headers,
json=data
)
print(response.json())

V. Ordinary User Tutorial for Using Gemini 3.1 Pro

Press enter or click to view image in full size

The configuration methods for these tools are basically similar — follow my steps to operate step by step, and you can use Gemini 3.1 Pro in 3–5 minutes. Most important reminder: The model names must be copied and pasted directly from Grsai’s official developer documentation (Chat API section) — do not type them manually! Even an extra space or wrong case will cause connection failure.

Press enter or click to view image in full size

1.Enter the settings page:

Download and install: Cherry Studio. After launching Cherry Studio, click the ‘Settings’ icon in the upper right corner (usually a gear shape).

2.Add model service: In the settings menu, find and click ‘Model Services,’ click Add, enter the provider name (custom, anything), and the model provider supports both OpenAI and Gemini.

3.Fill in provider information: Create ApiKey in Grsai console (https://grsai.com/dashboard/api-keys com domain requires VPN, change to ai), copy and paste.

Press enter or click to view image in full size

In Cherry Studio, paste ApiKey (keep confidential, do not share) and API address (base url/API Host): https://grsaiapi.com

Press enter or click to view image in full size

After filling in the Api key and address, manually add models by entering the model names supported in Grsai backend:

Press enter or click to view image in full size

Image generation models:

  • nano-banana-fast
  • nano-banana-pro

Chat models:

  • gemini-3-pro
  • gemini-3.1-pro
  • gemini-2.5-pro

After setup, return to the assistant page, select the model to use. After configuration, you can use Gemini 3.1 Pro in Cherry Studio. If you have bulk image generation needs, Cherry Studio is not suitable as it requires waiting for the conversation to end before continuing to ask questions — you can use Grsai’s open-source free bulk image generation tool.

1.Obtain APIKey: Grsai console settings ApiKey (https://grsai.com/zh/dashboard/api-keys com domain requires VPN, change to ai), copy ApiKey

Press enter or click to view image in full size

2.Configure API: Open the bulk image generation tool (https://image.grsai.com/ com domain requires VPN, change to ai), click settings in the upper right corner, paste your API Key

Press enter or click to view image in full size

3.Generate: Upload image, select model, enter size and prompt to use

Press enter or click to view image in full size

VI. Summary

评论

此博客中的热门博文

Configure Cherry Studio to Connect to Grsai in 5 Minutes, Achieving Freedom to Ask Questions and Draw Drawings with Gemini 3 Pro

Cheapest Sora 2 API in 2026: $0.08 Per Video + Full Tutorial & Character Consistency Guide

Stable AI API Proxy for Developers: GPT Image 1.5 via GrsAI ($0.02/Image)