Z.ai: GLM 4.5V

Public pricingIntelligence 79/100Medium memory视觉深度思考工具调用

Z.ai: GLM 4.5V 是一款多模态模型,适合智能体工作流与工具调用。它结合了多模态输入处理、可靠的工具调用与智能体表现、66K tokens上下文和均衡成本定位,可在agent workflows, tool use, and orchestration中提供可靠表现。它适合重视质量、速度与成本的团队,能带来稳定输出、灵活部署与扩展空间。 它适合需要稳定回答、较长上下文、清晰结构和可扩展部署的团队。 它适合需要稳定回答、较长上下文、清晰结构和可扩展部署的团队。 它适合需要稳定回答、较长上下文、清晰结构和可扩展部署的团队。 它适合需要稳定回答、较长上下文、清晰结构和可扩展部署的团队。 它适合需要稳定回答、较长上下文、清晰结构和可扩展部署的团队。 它适合需要稳定回答、较长上下文、清晰结构和可扩展部署的团队。

Input

$0.60/1M

Output

$1.80/1M

Cached

$0.11/1M

Batch

$0.30/1M

Calculate your GLM 4.5V bill.

Set your workload — see cost at your exact volume.

What would GLM 4.5V cost you?

Adjust the workload to see your monthly bill.

1,00010,00050,000250,0001M10M

Technical specifications

GLM 4.5V at a glance.

Memory

65,536

tokens

Max reply

16,384

tokens

Memory tier

Medium

a long report or a codebase file

Tokenizer

Released

Aug 2025

Training cutoff

Apr 2025

Availability

Public pricing

Status

active

What it can do

Capabilities & limits.

  • Understands images
  • Deep step-by-step thinking
  • Uses tools / calls functions
  • Strict JSON output
  • Streams replies
  • Fine-tunable on your data

When to pick GLM 4.5V

  • Multi-step reasoning, research agents, or hard math.
  • Screenshot analysis, image understanding, or document OCR.
  • Agentic workflows that call tools or APIs.
  • High-volume workloads where unit cost matters.

When to look elsewhere

  • Very latency-sensitive, real-time apps where every millisecond counts.

FAQ

GLM 4.5V — the questions we see most.

Pricing, capabilities, alternatives — generated from the same data that powers the calculator above.

Get instant answers from our AI agent

At a typical workload of 50,000 conversations a month with 1,500-token prompts and 800-token replies, GLM 4.5V costs roughly $117 per month. Input is $0.60 /1M tokens and output is $1.80 /1M tokens.
GLM 4.5V has a 65,536-token context window (medium memory — a long report or a codebase file). That means you can fit about 12,288 words of input and history in a single call.
Beyond text generation, GLM 4.5V supports understanding images, deep step-by-step reasoning, calling functions / tools, strict JSON output, fine-tuning on your own data. It streams replies by default.
GLM 4.5V was released in August 2025, with training data cut off around April 2025.
Models in a similar class include GLM 4.6V, Gemini 3 Flash Preview, GPT-5.4 Mini. The "Similar models" section below this FAQ links into each.

Still unsure?

Compare GLM 4.5V against 100+ other models.

Open the full wizard — pick a use case, set your usage, and see side-by-side monthly costs in under a minute.